var/home/core/zuul-output/0000755000175000017500000000000015114067310014523 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015114110770015466 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000006114051615114110761017676 0ustar rootrootDec 03 17:13:02 crc systemd[1]: Starting Kubernetes Kubelet... Dec 03 17:13:02 crc restorecon[4771]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:02 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:03 crc restorecon[4771]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 03 17:13:03 crc restorecon[4771]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 03 17:13:03 crc kubenswrapper[4787]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 17:13:03 crc kubenswrapper[4787]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 03 17:13:03 crc kubenswrapper[4787]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 17:13:03 crc kubenswrapper[4787]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 17:13:03 crc kubenswrapper[4787]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 03 17:13:03 crc kubenswrapper[4787]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.563372 4787 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566851 4787 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566888 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566895 4787 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566901 4787 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566906 4787 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566912 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566919 4787 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566926 4787 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566933 4787 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566939 4787 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566945 4787 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566950 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566955 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566959 4787 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566964 4787 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566968 4787 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566972 4787 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566977 4787 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566981 4787 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566984 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566989 4787 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566994 4787 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.566998 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567002 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567006 4787 feature_gate.go:330] unrecognized feature gate: Example Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567010 4787 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567014 4787 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567040 4787 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567052 4787 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567058 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567062 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567068 4787 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567073 4787 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567078 4787 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567082 4787 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567086 4787 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567092 4787 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567096 4787 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567101 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567106 4787 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567111 4787 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567117 4787 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567121 4787 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567126 4787 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567130 4787 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567134 4787 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567138 4787 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567142 4787 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567147 4787 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567151 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567156 4787 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567160 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567165 4787 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567171 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567175 4787 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567179 4787 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567184 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567188 4787 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567192 4787 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567200 4787 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567205 4787 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567209 4787 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567214 4787 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567218 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567222 4787 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567226 4787 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567231 4787 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567235 4787 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567239 4787 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567244 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.567248 4787 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567377 4787 flags.go:64] FLAG: --address="0.0.0.0" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567392 4787 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567403 4787 flags.go:64] FLAG: --anonymous-auth="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567410 4787 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567419 4787 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567426 4787 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567438 4787 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567448 4787 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567454 4787 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567460 4787 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567466 4787 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567471 4787 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567477 4787 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567483 4787 flags.go:64] FLAG: --cgroup-root="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567488 4787 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567494 4787 flags.go:64] FLAG: --client-ca-file="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567499 4787 flags.go:64] FLAG: --cloud-config="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567504 4787 flags.go:64] FLAG: --cloud-provider="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567510 4787 flags.go:64] FLAG: --cluster-dns="[]" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567518 4787 flags.go:64] FLAG: --cluster-domain="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567523 4787 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567529 4787 flags.go:64] FLAG: --config-dir="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567534 4787 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567540 4787 flags.go:64] FLAG: --container-log-max-files="5" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567549 4787 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567554 4787 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567559 4787 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567565 4787 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567570 4787 flags.go:64] FLAG: --contention-profiling="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567574 4787 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567579 4787 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567583 4787 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567588 4787 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567593 4787 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567597 4787 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567602 4787 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567606 4787 flags.go:64] FLAG: --enable-load-reader="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567610 4787 flags.go:64] FLAG: --enable-server="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567615 4787 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567621 4787 flags.go:64] FLAG: --event-burst="100" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567625 4787 flags.go:64] FLAG: --event-qps="50" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567629 4787 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567633 4787 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567639 4787 flags.go:64] FLAG: --eviction-hard="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567644 4787 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567649 4787 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567654 4787 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567658 4787 flags.go:64] FLAG: --eviction-soft="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567662 4787 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567667 4787 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567671 4787 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567675 4787 flags.go:64] FLAG: --experimental-mounter-path="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567680 4787 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567684 4787 flags.go:64] FLAG: --fail-swap-on="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567688 4787 flags.go:64] FLAG: --feature-gates="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567693 4787 flags.go:64] FLAG: --file-check-frequency="20s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567697 4787 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567701 4787 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567707 4787 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567712 4787 flags.go:64] FLAG: --healthz-port="10248" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567716 4787 flags.go:64] FLAG: --help="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567721 4787 flags.go:64] FLAG: --hostname-override="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567724 4787 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567729 4787 flags.go:64] FLAG: --http-check-frequency="20s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567733 4787 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567737 4787 flags.go:64] FLAG: --image-credential-provider-config="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567742 4787 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567746 4787 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567750 4787 flags.go:64] FLAG: --image-service-endpoint="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567754 4787 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567758 4787 flags.go:64] FLAG: --kube-api-burst="100" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567763 4787 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567767 4787 flags.go:64] FLAG: --kube-api-qps="50" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567771 4787 flags.go:64] FLAG: --kube-reserved="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567776 4787 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567779 4787 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567784 4787 flags.go:64] FLAG: --kubelet-cgroups="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567788 4787 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567792 4787 flags.go:64] FLAG: --lock-file="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567796 4787 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567801 4787 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567805 4787 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567811 4787 flags.go:64] FLAG: --log-json-split-stream="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567816 4787 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567820 4787 flags.go:64] FLAG: --log-text-split-stream="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567823 4787 flags.go:64] FLAG: --logging-format="text" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567828 4787 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567833 4787 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567837 4787 flags.go:64] FLAG: --manifest-url="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567841 4787 flags.go:64] FLAG: --manifest-url-header="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567847 4787 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567851 4787 flags.go:64] FLAG: --max-open-files="1000000" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567857 4787 flags.go:64] FLAG: --max-pods="110" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567861 4787 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567865 4787 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567869 4787 flags.go:64] FLAG: --memory-manager-policy="None" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567873 4787 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567878 4787 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567882 4787 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567887 4787 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567898 4787 flags.go:64] FLAG: --node-status-max-images="50" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567902 4787 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567906 4787 flags.go:64] FLAG: --oom-score-adj="-999" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567911 4787 flags.go:64] FLAG: --pod-cidr="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567915 4787 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567922 4787 flags.go:64] FLAG: --pod-manifest-path="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567927 4787 flags.go:64] FLAG: --pod-max-pids="-1" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567931 4787 flags.go:64] FLAG: --pods-per-core="0" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567935 4787 flags.go:64] FLAG: --port="10250" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567939 4787 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567944 4787 flags.go:64] FLAG: --provider-id="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567948 4787 flags.go:64] FLAG: --qos-reserved="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567952 4787 flags.go:64] FLAG: --read-only-port="10255" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567956 4787 flags.go:64] FLAG: --register-node="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567961 4787 flags.go:64] FLAG: --register-schedulable="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567965 4787 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567972 4787 flags.go:64] FLAG: --registry-burst="10" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567977 4787 flags.go:64] FLAG: --registry-qps="5" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567981 4787 flags.go:64] FLAG: --reserved-cpus="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567985 4787 flags.go:64] FLAG: --reserved-memory="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567995 4787 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.567999 4787 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568003 4787 flags.go:64] FLAG: --rotate-certificates="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568007 4787 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568011 4787 flags.go:64] FLAG: --runonce="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568033 4787 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568038 4787 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568043 4787 flags.go:64] FLAG: --seccomp-default="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568047 4787 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568051 4787 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568055 4787 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568060 4787 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568065 4787 flags.go:64] FLAG: --storage-driver-password="root" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568069 4787 flags.go:64] FLAG: --storage-driver-secure="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568074 4787 flags.go:64] FLAG: --storage-driver-table="stats" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568078 4787 flags.go:64] FLAG: --storage-driver-user="root" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568082 4787 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568086 4787 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568090 4787 flags.go:64] FLAG: --system-cgroups="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568095 4787 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568101 4787 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568105 4787 flags.go:64] FLAG: --tls-cert-file="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568110 4787 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568115 4787 flags.go:64] FLAG: --tls-min-version="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568119 4787 flags.go:64] FLAG: --tls-private-key-file="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568123 4787 flags.go:64] FLAG: --topology-manager-policy="none" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568127 4787 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568131 4787 flags.go:64] FLAG: --topology-manager-scope="container" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568135 4787 flags.go:64] FLAG: --v="2" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568141 4787 flags.go:64] FLAG: --version="false" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568147 4787 flags.go:64] FLAG: --vmodule="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568152 4787 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568158 4787 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568257 4787 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568263 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568268 4787 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568272 4787 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568276 4787 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568280 4787 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568284 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568288 4787 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568292 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568295 4787 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568300 4787 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568304 4787 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568307 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568312 4787 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568316 4787 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568320 4787 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568324 4787 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568328 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568332 4787 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568336 4787 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568340 4787 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568343 4787 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568347 4787 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568351 4787 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568354 4787 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568357 4787 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568361 4787 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568364 4787 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568368 4787 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568371 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568375 4787 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568380 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568383 4787 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568387 4787 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568391 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568394 4787 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568397 4787 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568401 4787 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568405 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568408 4787 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568411 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568415 4787 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568419 4787 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568423 4787 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568426 4787 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568429 4787 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568434 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568438 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568441 4787 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568445 4787 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568449 4787 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568453 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568457 4787 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568461 4787 feature_gate.go:330] unrecognized feature gate: Example Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568465 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568469 4787 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568472 4787 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568476 4787 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568480 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568483 4787 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568487 4787 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568490 4787 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568494 4787 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568498 4787 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568502 4787 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568506 4787 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568510 4787 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568513 4787 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568517 4787 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568522 4787 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.568527 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.568708 4787 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.576710 4787 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.576737 4787 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576805 4787 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576814 4787 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576819 4787 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576826 4787 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576832 4787 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576840 4787 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576845 4787 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576850 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576855 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576860 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576865 4787 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576870 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576874 4787 feature_gate.go:330] unrecognized feature gate: Example Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576878 4787 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576882 4787 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576886 4787 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576891 4787 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576895 4787 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576898 4787 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576903 4787 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576907 4787 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576911 4787 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576915 4787 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576920 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576925 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576930 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576936 4787 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576942 4787 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576947 4787 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576952 4787 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576957 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576962 4787 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576966 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576969 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576975 4787 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576979 4787 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576982 4787 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576986 4787 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576990 4787 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576995 4787 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.576999 4787 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577003 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577007 4787 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577011 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577038 4787 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577042 4787 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577046 4787 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577050 4787 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577053 4787 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577056 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577060 4787 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577063 4787 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577067 4787 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577071 4787 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577074 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577078 4787 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577081 4787 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577085 4787 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577089 4787 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577092 4787 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577096 4787 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577099 4787 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577103 4787 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577107 4787 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577112 4787 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577116 4787 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577119 4787 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577123 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577126 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577130 4787 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577134 4787 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.577141 4787 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577259 4787 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577272 4787 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577278 4787 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577285 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577290 4787 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577295 4787 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577299 4787 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577303 4787 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577307 4787 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577314 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577318 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577322 4787 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577326 4787 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577330 4787 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577334 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577338 4787 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577343 4787 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577348 4787 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577352 4787 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577357 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577362 4787 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577367 4787 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577371 4787 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577376 4787 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577382 4787 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577387 4787 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577391 4787 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577395 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577400 4787 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577404 4787 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577409 4787 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577413 4787 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577416 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577420 4787 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577424 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577430 4787 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577434 4787 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577438 4787 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577441 4787 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577445 4787 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577449 4787 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577453 4787 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577456 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577460 4787 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577463 4787 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577467 4787 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577471 4787 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577474 4787 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577477 4787 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577481 4787 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577486 4787 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577491 4787 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577497 4787 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577502 4787 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577506 4787 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577511 4787 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577516 4787 feature_gate.go:330] unrecognized feature gate: Example Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577520 4787 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577525 4787 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577529 4787 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577535 4787 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577539 4787 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577544 4787 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577549 4787 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577555 4787 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577559 4787 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577565 4787 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577569 4787 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577574 4787 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577579 4787 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.577585 4787 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.577593 4787 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.577921 4787 server.go:940] "Client rotation is on, will bootstrap in background" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.580687 4787 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.580764 4787 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.581229 4787 server.go:997] "Starting client certificate rotation" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.581250 4787 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.581465 4787 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-06 12:07:25.020438285 +0000 UTC Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.581595 4787 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 810h54m21.438846556s for next certificate rotation Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.600569 4787 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.602933 4787 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.624884 4787 log.go:25] "Validated CRI v1 runtime API" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.640411 4787 log.go:25] "Validated CRI v1 image API" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.642225 4787 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.645530 4787 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-03-17-08-22-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.645634 4787 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.671965 4787 manager.go:217] Machine: {Timestamp:2025-12-03 17:13:03.670925696 +0000 UTC m=+0.488396655 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:f359188a-c3b6-4925-b0ef-b6c8ba0e4e26 BootID:6d35e1b1-0d51-401a-9c96-4da3ec675199 Filesystems:[{Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:e4:a0:b0 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:e4:a0:b0 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:e0:d5:a3 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:e9:70:1f Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:c0:60:20 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:4b:a8:15 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:d7:12:9a Speed:-1 Mtu:1496} {Name:eth10 MacAddress:02:4c:a4:24:80:3d Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:a2:e8:67:9b:e6:36 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.672213 4787 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.672381 4787 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.672980 4787 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.673342 4787 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.673388 4787 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.673629 4787 topology_manager.go:138] "Creating topology manager with none policy" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.673642 4787 container_manager_linux.go:303] "Creating device plugin manager" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.673904 4787 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.673938 4787 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.674214 4787 state_mem.go:36] "Initialized new in-memory state store" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.674325 4787 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.680961 4787 kubelet.go:418] "Attempting to sync node with API server" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.680982 4787 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.681004 4787 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.681031 4787 kubelet.go:324] "Adding apiserver pod source" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.681046 4787 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.684464 4787 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.684890 4787 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.684953 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.684960 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.685040 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.685049 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.685650 4787 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686199 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686222 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686231 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686239 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686252 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686259 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686265 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686276 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686287 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686297 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686310 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.686318 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.694731 4787 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.695994 4787 server.go:1280] "Started kubelet" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.696092 4787 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.696236 4787 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.696865 4787 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.697486 4787 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 03 17:13:03 crc systemd[1]: Started Kubernetes Kubelet. Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.711804 4787 server.go:460] "Adding debug handlers to kubelet server" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.712672 4787 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.712746 4787 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.712798 4787 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-13 02:20:02.505994025 +0000 UTC Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.712824 4787 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.712840 4787 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.712887 4787 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.713064 4787 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.713454 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.713513 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.713637 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="200ms" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.714632 4787 factory.go:153] Registering CRI-O factory Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.714691 4787 factory.go:221] Registration of the crio container factory successfully Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.714806 4787 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.714856 4787 factory.go:55] Registering systemd factory Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.714870 4787 factory.go:221] Registration of the systemd container factory successfully Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.714898 4787 factory.go:103] Registering Raw factory Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.714950 4787 manager.go:1196] Started watching for new ooms in manager Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.714828 4787 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.65:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187dc3dd2c2e7675 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 17:13:03.695533685 +0000 UTC m=+0.513004664,LastTimestamp:2025-12-03 17:13:03.695533685 +0000 UTC m=+0.513004664,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.717436 4787 manager.go:319] Starting recovery of all containers Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728571 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728631 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728652 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728667 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728682 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728696 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728737 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728751 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728768 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728871 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728887 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728900 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728912 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728940 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728955 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728969 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.728982 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729116 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729126 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729135 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729202 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729219 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729232 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729243 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729257 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729266 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729281 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729292 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729304 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729313 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729324 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729338 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729348 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729363 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729372 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729382 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729398 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729408 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729419 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729429 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729438 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729448 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729462 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729473 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729485 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729495 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729504 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729514 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729546 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729557 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729568 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729579 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729593 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729605 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729614 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729625 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729635 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729646 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729657 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729667 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729677 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729688 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729703 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729714 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729724 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729736 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729746 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729757 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729768 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729778 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729789 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729799 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729810 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729819 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729832 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729842 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729851 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729866 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729876 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729886 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729896 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729912 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729922 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729939 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729948 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729964 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729975 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729986 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.729995 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.730009 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.730035 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.730051 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.730062 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.730072 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.730082 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.730091 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.730102 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.730460 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.731240 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.731382 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.731425 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.732174 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.732199 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.735594 4787 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.736704 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.736885 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.737003 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.737139 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.737227 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.737328 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.737414 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.737502 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.737583 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.737676 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.737760 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.737848 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.737928 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.738010 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.738141 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.738226 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.738512 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.738681 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.738793 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.738915 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.739073 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.739220 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.739345 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.739453 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.739577 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.739690 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.739820 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.739998 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.740155 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.740278 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.740408 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.740523 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.740638 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.740904 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.740988 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.741183 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.741309 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.741403 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.741482 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.741558 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.741643 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.741722 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.741810 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.741897 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.741992 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.742116 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.742226 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.742319 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.742406 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.742486 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.742591 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.742700 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.742825 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.742962 4787 manager.go:324] Recovery completed Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.742962 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743169 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743198 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743217 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743236 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743251 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743267 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743285 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743304 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743319 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743333 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743346 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743358 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743372 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743385 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743399 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743428 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743443 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743463 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743482 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743502 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743518 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743557 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743581 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743598 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743616 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743634 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743653 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743673 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743693 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743712 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743743 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743765 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743785 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743806 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743825 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743846 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743866 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743885 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743905 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743924 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743944 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743966 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.743986 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.744006 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.744048 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.744070 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.744089 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.744119 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.744138 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.744159 4787 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.744176 4787 reconstruct.go:97] "Volume reconstruction finished" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.744189 4787 reconciler.go:26] "Reconciler: start to sync state" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.753801 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.755355 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.755400 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.755416 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.756452 4787 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.756476 4787 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.756523 4787 state_mem.go:36] "Initialized new in-memory state store" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.759375 4787 policy_none.go:49] "None policy: Start" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.760079 4787 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.760106 4787 state_mem.go:35] "Initializing new in-memory state store" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.761851 4787 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.764623 4787 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.764677 4787 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.764718 4787 kubelet.go:2335] "Starting kubelet main sync loop" Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.764890 4787 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 03 17:13:03 crc kubenswrapper[4787]: W1203 17:13:03.766720 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.766785 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.813374 4787 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.816351 4787 manager.go:334] "Starting Device Plugin manager" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.816409 4787 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.816434 4787 server.go:79] "Starting device plugin registration server" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.817506 4787 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.817524 4787 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.817713 4787 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.817790 4787 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.817797 4787 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.826100 4787 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.865751 4787 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.865899 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.867311 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.867338 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.867351 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.867509 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.867728 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.867796 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.868672 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.868714 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.868731 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.868865 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.869002 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.869065 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.869076 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.869096 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.869121 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.870510 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.870921 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.870955 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.874484 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.874567 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.874586 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.874794 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.874862 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.874898 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.876010 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.876054 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.876065 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.876209 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.876633 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.876676 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.876660 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.876768 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.876690 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.877479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.877510 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.877524 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.877727 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.877756 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.879039 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.879072 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.879086 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.879346 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.879367 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.879378 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.914479 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="400ms" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.918495 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.919168 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.919192 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.919202 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.919219 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 17:13:03 crc kubenswrapper[4787]: E1203 17:13:03.919529 4787 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.65:6443: connect: connection refused" node="crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.946673 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.946731 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.946770 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.946840 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.946907 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.946978 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.947055 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.947100 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.947137 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.947221 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.947272 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.947329 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.947373 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.947422 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:03 crc kubenswrapper[4787]: I1203 17:13:03.947475 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.048810 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049323 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049353 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049379 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049389 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049124 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049405 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049480 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049483 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049538 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049500 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049511 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049591 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049629 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049685 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049708 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049734 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049732 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049745 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049696 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049826 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049830 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049743 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049756 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049890 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049910 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049921 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049932 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049952 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.049966 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.120311 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.126943 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.127001 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.127012 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.127060 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 17:13:04 crc kubenswrapper[4787]: E1203 17:13:04.127615 4787 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.65:6443: connect: connection refused" node="crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.213365 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.220229 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: W1203 17:13:04.245234 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-54d2fa760eb88b9bf09ec7063edbd042c370631cb8747f32827319fad9db2f7b WatchSource:0}: Error finding container 54d2fa760eb88b9bf09ec7063edbd042c370631cb8747f32827319fad9db2f7b: Status 404 returned error can't find the container with id 54d2fa760eb88b9bf09ec7063edbd042c370631cb8747f32827319fad9db2f7b Dec 03 17:13:04 crc kubenswrapper[4787]: W1203 17:13:04.247622 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-4daccf0b61b134d27ae6b7434eff8f0962a0acdb1d93e3c6da47ea05ba95450f WatchSource:0}: Error finding container 4daccf0b61b134d27ae6b7434eff8f0962a0acdb1d93e3c6da47ea05ba95450f: Status 404 returned error can't find the container with id 4daccf0b61b134d27ae6b7434eff8f0962a0acdb1d93e3c6da47ea05ba95450f Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.248496 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: W1203 17:13:04.260429 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-aa4918301685c1ff53a93bc0192bf266ecb7c9a3a4f1b74ef279a781b9eae331 WatchSource:0}: Error finding container aa4918301685c1ff53a93bc0192bf266ecb7c9a3a4f1b74ef279a781b9eae331: Status 404 returned error can't find the container with id aa4918301685c1ff53a93bc0192bf266ecb7c9a3a4f1b74ef279a781b9eae331 Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.269785 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.276653 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 03 17:13:04 crc kubenswrapper[4787]: W1203 17:13:04.289388 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-0c000abcd86f7cfc6e4a2bb8eb80b2234ad3f357943f63fe16070d30bf2ab1bd WatchSource:0}: Error finding container 0c000abcd86f7cfc6e4a2bb8eb80b2234ad3f357943f63fe16070d30bf2ab1bd: Status 404 returned error can't find the container with id 0c000abcd86f7cfc6e4a2bb8eb80b2234ad3f357943f63fe16070d30bf2ab1bd Dec 03 17:13:04 crc kubenswrapper[4787]: W1203 17:13:04.292559 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-5c7c123764512e8f5e15d3cab8b2374ccb8bbcb74a9dbc3454da20050588751c WatchSource:0}: Error finding container 5c7c123764512e8f5e15d3cab8b2374ccb8bbcb74a9dbc3454da20050588751c: Status 404 returned error can't find the container with id 5c7c123764512e8f5e15d3cab8b2374ccb8bbcb74a9dbc3454da20050588751c Dec 03 17:13:04 crc kubenswrapper[4787]: E1203 17:13:04.315105 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="800ms" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.528340 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.529845 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.529878 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.529888 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.529908 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 17:13:04 crc kubenswrapper[4787]: E1203 17:13:04.530298 4787 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.65:6443: connect: connection refused" node="crc" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.697808 4787 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.712905 4787 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2026-01-10 22:46:29.395938529 +0000 UTC Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.712957 4787 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 917h33m24.682983198s for next certificate rotation Dec 03 17:13:04 crc kubenswrapper[4787]: W1203 17:13:04.758955 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Dec 03 17:13:04 crc kubenswrapper[4787]: E1203 17:13:04.759067 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.776444 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb"} Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.776596 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"54d2fa760eb88b9bf09ec7063edbd042c370631cb8747f32827319fad9db2f7b"} Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.778359 4787 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1" exitCode=0 Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.778446 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1"} Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.778499 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"5c7c123764512e8f5e15d3cab8b2374ccb8bbcb74a9dbc3454da20050588751c"} Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.778865 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.780109 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.780149 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.780160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.781262 4787 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375" exitCode=0 Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.781321 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375"} Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.781342 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0c000abcd86f7cfc6e4a2bb8eb80b2234ad3f357943f63fe16070d30bf2ab1bd"} Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.781442 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.782549 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.782601 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.782616 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.783432 4787 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97" exitCode=0 Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.783501 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97"} Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.783524 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"aa4918301685c1ff53a93bc0192bf266ecb7c9a3a4f1b74ef279a781b9eae331"} Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.783585 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.784339 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.784372 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.784386 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.785176 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.785792 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.785814 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.785823 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.786223 4787 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd" exitCode=0 Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.786253 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd"} Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.786274 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"4daccf0b61b134d27ae6b7434eff8f0962a0acdb1d93e3c6da47ea05ba95450f"} Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.786365 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.786993 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.787064 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:04 crc kubenswrapper[4787]: I1203 17:13:04.787074 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:04 crc kubenswrapper[4787]: W1203 17:13:04.874347 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Dec 03 17:13:04 crc kubenswrapper[4787]: E1203 17:13:04.874468 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Dec 03 17:13:04 crc kubenswrapper[4787]: W1203 17:13:04.898811 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Dec 03 17:13:04 crc kubenswrapper[4787]: E1203 17:13:04.898920 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Dec 03 17:13:05 crc kubenswrapper[4787]: E1203 17:13:05.115880 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="1.6s" Dec 03 17:13:05 crc kubenswrapper[4787]: W1203 17:13:05.237894 4787 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Dec 03 17:13:05 crc kubenswrapper[4787]: E1203 17:13:05.237992 4787 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.65:6443: connect: connection refused" logger="UnhandledError" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.331106 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.332703 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.332750 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.332761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.332791 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 17:13:05 crc kubenswrapper[4787]: E1203 17:13:05.333496 4787 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.65:6443: connect: connection refused" node="crc" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.701886 4787 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.65:6443: connect: connection refused Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.794752 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.794830 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.794844 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.794856 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.794868 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.794986 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.795748 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.795778 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.795789 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.796748 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"80c7d63868c2a5ff22d609f1c3bb94a939eca71443e7b274c77219c4b641e12d"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.796856 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.797725 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.797762 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.797776 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.798996 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.799057 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.799072 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.799140 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.799733 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.799760 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.799771 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.801100 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.801127 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.801139 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.801141 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.801957 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.801983 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.801993 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.802746 4787 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb" exitCode=0 Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.802777 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb"} Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.802895 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.803511 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.803543 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:05 crc kubenswrapper[4787]: I1203 17:13:05.803553 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.143834 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.638130 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.809482 4787 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b" exitCode=0 Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.809540 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b"} Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.809662 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.809682 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.809738 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.811680 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.811726 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.811740 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.811752 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.811761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.811686 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.811845 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.811875 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.811767 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.934216 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.935468 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.935529 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.935547 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:06 crc kubenswrapper[4787]: I1203 17:13:06.935585 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 17:13:07 crc kubenswrapper[4787]: I1203 17:13:07.821667 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:07 crc kubenswrapper[4787]: I1203 17:13:07.821535 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e"} Dec 03 17:13:07 crc kubenswrapper[4787]: I1203 17:13:07.821836 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf"} Dec 03 17:13:07 crc kubenswrapper[4787]: I1203 17:13:07.821874 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf"} Dec 03 17:13:07 crc kubenswrapper[4787]: I1203 17:13:07.822133 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b"} Dec 03 17:13:07 crc kubenswrapper[4787]: I1203 17:13:07.823168 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:07 crc kubenswrapper[4787]: I1203 17:13:07.823237 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:07 crc kubenswrapper[4787]: I1203 17:13:07.823268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:08 crc kubenswrapper[4787]: I1203 17:13:08.643644 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:08 crc kubenswrapper[4787]: I1203 17:13:08.643867 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:08 crc kubenswrapper[4787]: I1203 17:13:08.645273 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:08 crc kubenswrapper[4787]: I1203 17:13:08.645319 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:08 crc kubenswrapper[4787]: I1203 17:13:08.645335 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:08 crc kubenswrapper[4787]: I1203 17:13:08.828488 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66"} Dec 03 17:13:08 crc kubenswrapper[4787]: I1203 17:13:08.829714 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:08 crc kubenswrapper[4787]: I1203 17:13:08.831277 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:08 crc kubenswrapper[4787]: I1203 17:13:08.831327 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:08 crc kubenswrapper[4787]: I1203 17:13:08.831337 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.002733 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.492823 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.492995 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.494172 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.494332 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.494533 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.499211 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.540579 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.540863 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.542299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.542394 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.542414 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.831161 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.831161 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.831967 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.831999 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.832008 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.832500 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.832519 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:09 crc kubenswrapper[4787]: I1203 17:13:09.832543 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:10 crc kubenswrapper[4787]: I1203 17:13:10.324387 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 17:13:10 crc kubenswrapper[4787]: I1203 17:13:10.324604 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:10 crc kubenswrapper[4787]: I1203 17:13:10.326147 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:10 crc kubenswrapper[4787]: I1203 17:13:10.326178 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:10 crc kubenswrapper[4787]: I1203 17:13:10.326186 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:10 crc kubenswrapper[4787]: I1203 17:13:10.834643 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:10 crc kubenswrapper[4787]: I1203 17:13:10.835788 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:10 crc kubenswrapper[4787]: I1203 17:13:10.835855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:10 crc kubenswrapper[4787]: I1203 17:13:10.835881 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:11 crc kubenswrapper[4787]: I1203 17:13:11.368457 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:11 crc kubenswrapper[4787]: I1203 17:13:11.368656 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 17:13:11 crc kubenswrapper[4787]: I1203 17:13:11.368719 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:11 crc kubenswrapper[4787]: I1203 17:13:11.369955 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:11 crc kubenswrapper[4787]: I1203 17:13:11.369980 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:11 crc kubenswrapper[4787]: I1203 17:13:11.369989 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:13 crc kubenswrapper[4787]: I1203 17:13:13.413703 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:13 crc kubenswrapper[4787]: I1203 17:13:13.413993 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:13 crc kubenswrapper[4787]: I1203 17:13:13.415564 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:13 crc kubenswrapper[4787]: I1203 17:13:13.415612 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:13 crc kubenswrapper[4787]: I1203 17:13:13.415623 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:13 crc kubenswrapper[4787]: E1203 17:13:13.826264 4787 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 03 17:13:14 crc kubenswrapper[4787]: I1203 17:13:14.368633 4787 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 17:13:14 crc kubenswrapper[4787]: I1203 17:13:14.368738 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 17:13:14 crc kubenswrapper[4787]: I1203 17:13:14.842828 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 03 17:13:14 crc kubenswrapper[4787]: I1203 17:13:14.843096 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:14 crc kubenswrapper[4787]: I1203 17:13:14.844697 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:14 crc kubenswrapper[4787]: I1203 17:13:14.844756 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:14 crc kubenswrapper[4787]: I1203 17:13:14.844768 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:16 crc kubenswrapper[4787]: I1203 17:13:16.699518 4787 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Dec 03 17:13:16 crc kubenswrapper[4787]: E1203 17:13:16.716884 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" interval="3.2s" Dec 03 17:13:16 crc kubenswrapper[4787]: I1203 17:13:16.908195 4787 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 17:13:16 crc kubenswrapper[4787]: I1203 17:13:16.908294 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 17:13:16 crc kubenswrapper[4787]: I1203 17:13:16.914339 4787 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 03 17:13:16 crc kubenswrapper[4787]: I1203 17:13:16.914412 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 03 17:13:19 crc kubenswrapper[4787]: I1203 17:13:19.546884 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:19 crc kubenswrapper[4787]: I1203 17:13:19.547066 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:19 crc kubenswrapper[4787]: I1203 17:13:19.548569 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:19 crc kubenswrapper[4787]: I1203 17:13:19.548622 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:19 crc kubenswrapper[4787]: I1203 17:13:19.548632 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:19 crc kubenswrapper[4787]: I1203 17:13:19.554900 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:19 crc kubenswrapper[4787]: I1203 17:13:19.858283 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 17:13:19 crc kubenswrapper[4787]: I1203 17:13:19.858359 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:19 crc kubenswrapper[4787]: I1203 17:13:19.859464 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:19 crc kubenswrapper[4787]: I1203 17:13:19.859506 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:19 crc kubenswrapper[4787]: I1203 17:13:19.859523 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.887482 4787 trace.go:236] Trace[923560955]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 17:13:06.905) (total time: 14982ms): Dec 03 17:13:21 crc kubenswrapper[4787]: Trace[923560955]: ---"Objects listed" error: 14982ms (17:13:21.887) Dec 03 17:13:21 crc kubenswrapper[4787]: Trace[923560955]: [14.982055966s] [14.982055966s] END Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.887863 4787 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.888089 4787 trace.go:236] Trace[76893634]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 17:13:06.876) (total time: 15011ms): Dec 03 17:13:21 crc kubenswrapper[4787]: Trace[76893634]: ---"Objects listed" error: 15011ms (17:13:21.887) Dec 03 17:13:21 crc kubenswrapper[4787]: Trace[76893634]: [15.011894877s] [15.011894877s] END Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.888129 4787 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.889537 4787 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 03 17:13:21 crc kubenswrapper[4787]: E1203 17:13:21.890057 4787 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.890282 4787 trace.go:236] Trace[527512167]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 17:13:07.176) (total time: 14713ms): Dec 03 17:13:21 crc kubenswrapper[4787]: Trace[527512167]: ---"Objects listed" error: 14713ms (17:13:21.890) Dec 03 17:13:21 crc kubenswrapper[4787]: Trace[527512167]: [14.713845256s] [14.713845256s] END Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.890335 4787 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.894838 4787 trace.go:236] Trace[1160866422]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (03-Dec-2025 17:13:08.239) (total time: 13655ms): Dec 03 17:13:21 crc kubenswrapper[4787]: Trace[1160866422]: ---"Objects listed" error: 13654ms (17:13:21.894) Dec 03 17:13:21 crc kubenswrapper[4787]: Trace[1160866422]: [13.655079198s] [13.655079198s] END Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.894886 4787 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.937462 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.945685 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.954110 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.961677 4787 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36196->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.961734 4787 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36212->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.961776 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36196->192.168.126.11:17697: read: connection reset by peer" Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.961802 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:36212->192.168.126.11:17697: read: connection reset by peer" Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.962291 4787 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 03 17:13:21 crc kubenswrapper[4787]: I1203 17:13:21.962376 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.695476 4787 apiserver.go:52] "Watching apiserver" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.698533 4787 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.698988 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-5rllg","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-machine-config-operator/machine-config-daemon-t882k","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.699427 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.699554 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.699745 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.699750 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.699899 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.699961 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.700127 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.700115 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.700188 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.700415 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5rllg" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.700421 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.700977 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.702380 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.702427 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.702463 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.702518 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.702897 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.703310 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.703660 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.703757 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.703869 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.704293 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.704495 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.704572 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.704651 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.704708 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.704729 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.704977 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.713843 4787 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.714624 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.724655 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.736365 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.749185 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.761471 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.775423 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.784896 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794065 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794118 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794145 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794174 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794198 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794220 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794246 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794267 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794296 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794326 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794355 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794383 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794413 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794445 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794473 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794505 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794533 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794563 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794608 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794640 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794671 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794702 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794731 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794737 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794764 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794797 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794831 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794865 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794897 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794929 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794956 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794984 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795041 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795075 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795107 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795143 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795177 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795213 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795244 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795297 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795328 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795355 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795383 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795408 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795438 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795522 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795557 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795585 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795612 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795639 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795668 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795738 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795767 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796085 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796128 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796216 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796331 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796371 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796411 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796451 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800195 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800260 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800362 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800455 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800498 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800580 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800634 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800696 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800799 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800843 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800975 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.801226 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.801344 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.801393 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.801436 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.801535 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.801603 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.801690 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.801785 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.801909 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.802302 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.802389 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.802442 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794822 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.794841 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.802565 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795227 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795258 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.802629 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795741 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795767 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795783 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795823 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795784 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795862 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795947 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795965 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796262 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796253 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796282 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796302 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796451 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.797964 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.798117 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.798140 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.796506 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.798432 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.798565 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.798577 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.798664 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.798678 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.798901 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.798967 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.799240 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.799417 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.799506 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.799835 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.799748 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800160 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.800467 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.801500 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.802079 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.802237 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.795552 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.802645 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.803203 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.803257 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.803419 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.803465 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.803708 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.803738 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.804845 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.804941 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.805202 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.805513 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.805884 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.807107 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.807245 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.804204 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.807497 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.807503 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.807652 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.807663 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.807690 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.807750 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.807801 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.807841 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.807966 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808031 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808164 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808265 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808363 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808721 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808832 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808915 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809029 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809115 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809195 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809297 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809383 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809458 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809530 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809607 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809687 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809765 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809844 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809919 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810079 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809278 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810809 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.813076 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.813922 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.814029 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.814072 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.814107 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.814144 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.814175 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.814208 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.814239 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.814291 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808385 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808504 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808527 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808618 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808690 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.808802 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809074 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809080 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809232 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809303 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809347 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.803429 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.814541 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.809969 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810354 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810403 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810550 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810732 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810801 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810912 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810929 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810910 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810947 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.811035 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.811100 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.811763 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.811835 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.811847 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.812473 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.810951 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.813233 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.814121 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.815172 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.815313 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.815670 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.815866 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.815939 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.815980 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.816010 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.816085 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.816090 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.816339 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.816353 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.816496 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.816507 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.816563 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.816651 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.816685 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.816804 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.817767 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.818588 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.818954 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.819110 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.820612 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.820646 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.820750 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.820881 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821171 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821153 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821218 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821356 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821362 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821378 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821413 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821452 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821485 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821535 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821710 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.821857 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822193 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822384 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822395 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822426 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822450 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822473 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822497 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822519 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822543 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822564 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822585 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822607 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822630 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822651 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822674 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822695 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.823120 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.823157 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.823181 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822448 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.823205 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.823232 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822599 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822730 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822862 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.822887 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.823192 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.823365 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.823255 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824157 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824190 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824218 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824242 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824265 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824288 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824311 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824344 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824366 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824388 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824411 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824434 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824456 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824478 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824498 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824522 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824546 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824569 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824594 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824616 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824639 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824662 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824696 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824723 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826434 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826488 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826514 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826600 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826628 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826648 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826669 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826688 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826713 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826738 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826772 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826791 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826809 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826828 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826965 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826990 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827006 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827045 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827070 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827095 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827119 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827169 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827195 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2dfr\" (UniqueName: \"kubernetes.io/projected/1f5d3d8c-cdec-4ed3-bb53-0a5012751616-kube-api-access-l2dfr\") pod \"node-resolver-5rllg\" (UID: \"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\") " pod="openshift-dns/node-resolver-5rllg" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827215 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b6597ac6-3ab2-4d2f-b38e-896795a7773d-mcd-auth-proxy-config\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827232 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-df4v9\" (UniqueName: \"kubernetes.io/projected/b6597ac6-3ab2-4d2f-b38e-896795a7773d-kube-api-access-df4v9\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827251 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827270 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827288 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827325 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1f5d3d8c-cdec-4ed3-bb53-0a5012751616-hosts-file\") pod \"node-resolver-5rllg\" (UID: \"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\") " pod="openshift-dns/node-resolver-5rllg" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827343 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827361 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827380 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827398 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827415 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827433 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827456 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827483 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827512 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b6597ac6-3ab2-4d2f-b38e-896795a7773d-proxy-tls\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827537 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827556 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b6597ac6-3ab2-4d2f-b38e-896795a7773d-rootfs\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827578 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824915 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827722 4787 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827742 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827754 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827764 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827774 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827784 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827793 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827805 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827817 4787 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827831 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827842 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827855 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827867 4787 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827879 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827892 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827903 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827913 4787 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827922 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827948 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827957 4787 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827972 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827982 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827990 4787 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828001 4787 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828011 4787 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828039 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828133 4787 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828150 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828162 4787 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828174 4787 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828195 4787 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828208 4787 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.829422 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824904 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.824933 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.825406 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.825406 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.825426 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.825728 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.825789 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.825962 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826117 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826207 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826221 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826247 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.826493 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827617 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.827668 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828114 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.828255 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:13:23.328216054 +0000 UTC m=+20.145687013 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.829633 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828453 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828480 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828666 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.828725 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.829349 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.829669 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.829763 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:23.329744525 +0000 UTC m=+20.147215484 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.829801 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.829912 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.830046 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.830056 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.830083 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:23.330074724 +0000 UTC m=+20.147545683 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.829917 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.829959 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.830216 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.831030 4787 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.831677 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832120 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832186 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832418 4787 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832444 4787 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832455 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832467 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832477 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832488 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832498 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832510 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832519 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832529 4787 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832540 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832549 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832558 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832567 4787 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832577 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832586 4787 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832596 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832606 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832636 4787 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832649 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832658 4787 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832668 4787 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832659 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832678 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832738 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832763 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832786 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832812 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832836 4787 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832855 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832875 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832896 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832916 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832936 4787 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832938 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832958 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832977 4787 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.832996 4787 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833037 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833058 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833076 4787 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833093 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833111 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833128 4787 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833147 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833177 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833590 4787 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833611 4787 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833630 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833651 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833666 4787 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833684 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833701 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833719 4787 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833736 4787 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833581 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833758 4787 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833778 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833796 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833815 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833833 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833854 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833873 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833892 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833910 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833724 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833856 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.833942 4787 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834001 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834036 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834067 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834083 4787 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834096 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834111 4787 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834123 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834139 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834154 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834152 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834169 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834182 4787 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834195 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834210 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834222 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834241 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834251 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834261 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834271 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834282 4787 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834291 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834300 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834312 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834322 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834332 4787 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834342 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834351 4787 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834361 4787 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834370 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834380 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834389 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834398 4787 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.834407 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.835003 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.836553 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.837514 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.837537 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.837911 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.838078 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.838165 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.838220 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.838261 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.838351 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.838528 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.839088 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.839544 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.840259 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.840865 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.841628 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.846473 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.846922 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.847116 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.850812 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.853245 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.856381 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.857619 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.860241 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.860650 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.860689 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.860704 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.860712 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.860763 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:23.360744897 +0000 UTC m=+20.178215846 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.861110 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.861156 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.861626 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.861869 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.861868 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.864053 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.871755 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.872709 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.874399 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.875529 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.876163 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.876512 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.876543 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.876558 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.876629 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:23.376605017 +0000 UTC m=+20.194076196 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.877474 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.881370 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.886975 4787 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb" exitCode=255 Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.887221 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb"} Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.888676 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-vj6t6"] Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.889453 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.891002 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-xcz4f"] Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.891399 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.894567 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.894604 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.894819 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.894857 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bx8nf"] Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.894944 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.895067 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.895088 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.895249 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.896678 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: E1203 17:13:22.896852 4787 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.901622 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.901829 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.902007 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.902333 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.902477 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.902614 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.902815 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.906786 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.909421 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.922386 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.934779 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-var-lib-cni-bin\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.935052 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-os-release\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.935184 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-system-cni-dir\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.935286 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-kubelet\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.935384 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-bin\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.935494 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdlvg\" (UniqueName: \"kubernetes.io/projected/bd5617c0-aa4f-4f21-a131-15af831725ec-kube-api-access-tdlvg\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.935588 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-var-lib-kubelet\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.935686 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-hostroot\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.935822 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b6597ac6-3ab2-4d2f-b38e-896795a7773d-proxy-tls\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.935917 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.936040 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/34f0f91e-2de6-4e13-974c-917d871c3a43-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.936142 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b6597ac6-3ab2-4d2f-b38e-896795a7773d-mcd-auth-proxy-config\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.936254 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2dfr\" (UniqueName: \"kubernetes.io/projected/1f5d3d8c-cdec-4ed3-bb53-0a5012751616-kube-api-access-l2dfr\") pod \"node-resolver-5rllg\" (UID: \"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\") " pod="openshift-dns/node-resolver-5rllg" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.936349 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-cnibin\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.936443 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-node-log\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.936534 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-env-overrides\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.936633 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-log-socket\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.936744 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1f5d3d8c-cdec-4ed3-bb53-0a5012751616-hosts-file\") pod \"node-resolver-5rllg\" (UID: \"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\") " pod="openshift-dns/node-resolver-5rllg" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.936844 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-system-cni-dir\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.936931 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-etc-openvswitch\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.937031 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-ovn\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.937134 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.936059 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.937389 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-ovn-kubernetes\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.937507 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-cnibin\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.937612 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/34f0f91e-2de6-4e13-974c-917d871c3a43-cni-binary-copy\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.937709 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-openvswitch\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.937816 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-netd\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.937921 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5k4bd\" (UniqueName: \"kubernetes.io/projected/c3f5a989-6421-4bc7-a516-b37c825d8bbc-kube-api-access-5k4bd\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938041 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-daemon-config\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938166 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.937182 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938278 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/b6597ac6-3ab2-4d2f-b38e-896795a7773d-mcd-auth-proxy-config\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938265 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938345 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rrl8n\" (UniqueName: \"kubernetes.io/projected/34f0f91e-2de6-4e13-974c-917d871c3a43-kube-api-access-rrl8n\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938388 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-cni-dir\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938421 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-socket-dir-parent\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938448 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-var-lib-openvswitch\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938578 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/1f5d3d8c-cdec-4ed3-bb53-0a5012751616-hosts-file\") pod \"node-resolver-5rllg\" (UID: \"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\") " pod="openshift-dns/node-resolver-5rllg" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938629 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-config\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938682 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b6597ac6-3ab2-4d2f-b38e-896795a7773d-rootfs\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938706 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-slash\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938730 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-script-lib\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938797 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/b6597ac6-3ab2-4d2f-b38e-896795a7773d-rootfs\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938781 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938818 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-df4v9\" (UniqueName: \"kubernetes.io/projected/b6597ac6-3ab2-4d2f-b38e-896795a7773d-kube-api-access-df4v9\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938872 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-os-release\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938926 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-run-netns\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938949 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-run-multus-certs\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938973 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-etc-kubernetes\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.938997 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-systemd-units\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939043 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-netns\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939064 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovn-node-metrics-cert\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939095 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bd5617c0-aa4f-4f21-a131-15af831725ec-cni-binary-copy\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939121 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-run-k8s-cni-cncf-io\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939140 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-var-lib-cni-multus\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939163 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-conf-dir\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939167 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/b6597ac6-3ab2-4d2f-b38e-896795a7773d-proxy-tls\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939201 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-systemd\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939585 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939687 4787 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939773 4787 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939853 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.939930 4787 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940155 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940243 4787 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940328 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940404 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940485 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940546 4787 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940608 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940663 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940716 4787 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940773 4787 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940825 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940883 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.940955 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.941063 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.941149 4787 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.941236 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.941344 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.941432 4787 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.941516 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.941603 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.941685 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.941765 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.941852 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.941941 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942049 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942143 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942220 4787 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942293 4787 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942381 4787 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942462 4787 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942545 4787 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942632 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942711 4787 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942784 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942865 4787 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.942947 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943039 4787 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943133 4787 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943202 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943285 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943374 4787 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943454 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943529 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943619 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943697 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943782 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943871 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.943953 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944082 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944170 4787 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944233 4787 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944291 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944347 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944408 4787 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944485 4787 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944563 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944645 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944725 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944815 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.944915 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.945004 4787 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.945120 4787 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.960071 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.960543 4787 scope.go:117] "RemoveContainer" containerID="13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.961410 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2dfr\" (UniqueName: \"kubernetes.io/projected/1f5d3d8c-cdec-4ed3-bb53-0a5012751616-kube-api-access-l2dfr\") pod \"node-resolver-5rllg\" (UID: \"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\") " pod="openshift-dns/node-resolver-5rllg" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.964804 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.968290 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-df4v9\" (UniqueName: \"kubernetes.io/projected/b6597ac6-3ab2-4d2f-b38e-896795a7773d-kube-api-access-df4v9\") pod \"machine-config-daemon-t882k\" (UID: \"b6597ac6-3ab2-4d2f-b38e-896795a7773d\") " pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:22 crc kubenswrapper[4787]: I1203 17:13:22.979600 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.013215 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.020839 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.024490 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.032534 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.035900 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-5rllg" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.043843 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.046787 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-var-lib-kubelet\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.046828 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-hostroot\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.046856 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdlvg\" (UniqueName: \"kubernetes.io/projected/bd5617c0-aa4f-4f21-a131-15af831725ec-kube-api-access-tdlvg\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.046884 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/34f0f91e-2de6-4e13-974c-917d871c3a43-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.046908 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-cnibin\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.046930 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-node-log\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.046949 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-env-overrides\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.046968 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-log-socket\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.046987 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-system-cni-dir\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047038 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-etc-openvswitch\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047061 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-ovn\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047083 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047118 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-ovn-kubernetes\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047152 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/34f0f91e-2de6-4e13-974c-917d871c3a43-cni-binary-copy\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047174 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-openvswitch\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047191 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-netd\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047211 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5k4bd\" (UniqueName: \"kubernetes.io/projected/c3f5a989-6421-4bc7-a516-b37c825d8bbc-kube-api-access-5k4bd\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047242 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-cnibin\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047262 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047283 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rrl8n\" (UniqueName: \"kubernetes.io/projected/34f0f91e-2de6-4e13-974c-917d871c3a43-kube-api-access-rrl8n\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047304 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-cni-dir\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047325 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-socket-dir-parent\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047346 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-daemon-config\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047367 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-var-lib-openvswitch\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047384 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-config\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047417 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-slash\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047437 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-run-multus-certs\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047457 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-etc-kubernetes\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047478 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-systemd-units\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047498 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-netns\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047519 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-script-lib\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047540 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-os-release\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047560 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-run-netns\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047583 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-run-k8s-cni-cncf-io\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047601 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-var-lib-cni-multus\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047619 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-conf-dir\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047640 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-systemd\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047659 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovn-node-metrics-cert\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047679 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bd5617c0-aa4f-4f21-a131-15af831725ec-cni-binary-copy\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047697 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-var-lib-cni-bin\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047719 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-kubelet\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047737 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-bin\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047757 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-os-release\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047777 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-system-cni-dir\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047865 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-system-cni-dir\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047913 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-var-lib-kubelet\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.047941 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-hostroot\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.048259 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-var-lib-openvswitch\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.048341 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-ovn-kubernetes\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049035 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-daemon-config\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049150 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-run-k8s-cni-cncf-io\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049224 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-systemd-units\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049240 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-ovn\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049259 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-slash\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049255 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-netns\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049304 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-run-multus-certs\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049315 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049344 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-etc-kubernetes\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049351 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-log-socket\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049498 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-cni-dir\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049485 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-os-release\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049554 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-os-release\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049582 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-run-netns\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049598 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-cnibin\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049927 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/34f0f91e-2de6-4e13-974c-917d871c3a43-cni-binary-copy\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.049982 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-socket-dir-parent\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050008 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-openvswitch\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050010 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-script-lib\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050052 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-netd\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050539 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/bd5617c0-aa4f-4f21-a131-15af831725ec-cni-binary-copy\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050625 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/34f0f91e-2de6-4e13-974c-917d871c3a43-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050659 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-var-lib-cni-bin\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050685 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-system-cni-dir\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050687 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-bin\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050701 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-multus-conf-dir\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050702 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-node-log\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050719 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-etc-openvswitch\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050732 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-systemd\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050757 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-cnibin\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050771 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-kubelet\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.051066 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/34f0f91e-2de6-4e13-974c-917d871c3a43-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.050734 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/bd5617c0-aa4f-4f21-a131-15af831725ec-host-var-lib-cni-multus\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.053575 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-env-overrides\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.055759 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-config\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.056513 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovn-node-metrics-cert\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.063446 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.070145 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdlvg\" (UniqueName: \"kubernetes.io/projected/bd5617c0-aa4f-4f21-a131-15af831725ec-kube-api-access-tdlvg\") pod \"multus-xcz4f\" (UID: \"bd5617c0-aa4f-4f21-a131-15af831725ec\") " pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.074045 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5k4bd\" (UniqueName: \"kubernetes.io/projected/c3f5a989-6421-4bc7-a516-b37c825d8bbc-kube-api-access-5k4bd\") pod \"ovnkube-node-bx8nf\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.079878 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rrl8n\" (UniqueName: \"kubernetes.io/projected/34f0f91e-2de6-4e13-974c-917d871c3a43-kube-api-access-rrl8n\") pod \"multus-additional-cni-plugins-vj6t6\" (UID: \"34f0f91e-2de6-4e13-974c-917d871c3a43\") " pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.092549 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.123123 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.136829 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: W1203 17:13:23.142208 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb6597ac6_3ab2_4d2f_b38e_896795a7773d.slice/crio-176783ac6fbfceb6c309e97b1cfd5a47e1fff2fae18e0a99e4e18c43b7a71b1d WatchSource:0}: Error finding container 176783ac6fbfceb6c309e97b1cfd5a47e1fff2fae18e0a99e4e18c43b7a71b1d: Status 404 returned error can't find the container with id 176783ac6fbfceb6c309e97b1cfd5a47e1fff2fae18e0a99e4e18c43b7a71b1d Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.148668 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.168934 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.181871 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.194971 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.210950 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.214997 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.220588 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-xcz4f" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.222443 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.232264 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.234443 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: W1203 17:13:23.243745 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34f0f91e_2de6_4e13_974c_917d871c3a43.slice/crio-df3b45024336a5d8fa0fafae1bcbb491dac2ea23d6283d3d07dc67b388091b61 WatchSource:0}: Error finding container df3b45024336a5d8fa0fafae1bcbb491dac2ea23d6283d3d07dc67b388091b61: Status 404 returned error can't find the container with id df3b45024336a5d8fa0fafae1bcbb491dac2ea23d6283d3d07dc67b388091b61 Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.250085 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: W1203 17:13:23.254257 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbd5617c0_aa4f_4f21_a131_15af831725ec.slice/crio-20da6a02178ade653848354773e1e7a7414604bbd5326e50d4f9832de52f2317 WatchSource:0}: Error finding container 20da6a02178ade653848354773e1e7a7414604bbd5326e50d4f9832de52f2317: Status 404 returned error can't find the container with id 20da6a02178ade653848354773e1e7a7414604bbd5326e50d4f9832de52f2317 Dec 03 17:13:23 crc kubenswrapper[4787]: W1203 17:13:23.256619 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc3f5a989_6421_4bc7_a516_b37c825d8bbc.slice/crio-3c022f138769ab1d64f295dda0b5d503db70d70899c44ff8fa396d4de1d9824d WatchSource:0}: Error finding container 3c022f138769ab1d64f295dda0b5d503db70d70899c44ff8fa396d4de1d9824d: Status 404 returned error can't find the container with id 3c022f138769ab1d64f295dda0b5d503db70d70899c44ff8fa396d4de1d9824d Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.262860 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.274355 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.285887 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.304567 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.350777 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.350904 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.350951 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.351041 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.351088 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:24.351076002 +0000 UTC m=+21.168546961 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.351436 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:13:24.351427151 +0000 UTC m=+21.168898110 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.351505 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.351531 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:24.351524294 +0000 UTC m=+21.168995253 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.451524 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.451593 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.451706 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.451722 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.451733 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.451729 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.451763 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.451775 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.451781 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:24.451768351 +0000 UTC m=+21.269239310 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:23 crc kubenswrapper[4787]: E1203 17:13:23.451829 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:24.451813912 +0000 UTC m=+21.269284871 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.771279 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.772176 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.773541 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.774404 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.776014 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.776731 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.777590 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.778708 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.779655 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.780660 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.781281 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.782745 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.783461 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.784208 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.784592 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.785714 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.786312 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.787624 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.788258 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.788876 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.790098 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.790776 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.791997 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.792506 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.793581 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.794083 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.794695 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.795758 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.796274 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.797247 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.797803 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.798704 4787 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.798867 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.800644 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.801910 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.802426 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.802895 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.804231 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.804956 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.806006 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.806666 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.808083 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.808690 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.811302 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.811946 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.813037 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.813587 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.814617 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.815334 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.816454 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.817131 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.818168 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.819351 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.820040 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.820735 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.821536 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.822676 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.846354 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.866746 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.882529 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.891116 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xcz4f" event={"ID":"bd5617c0-aa4f-4f21-a131-15af831725ec","Type":"ContainerStarted","Data":"d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.892057 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xcz4f" event={"ID":"bd5617c0-aa4f-4f21-a131-15af831725ec","Type":"ContainerStarted","Data":"20da6a02178ade653848354773e1e7a7414604bbd5326e50d4f9832de52f2317"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.897065 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.901505 4787 generic.go:334] "Generic (PLEG): container finished" podID="34f0f91e-2de6-4e13-974c-917d871c3a43" containerID="8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe" exitCode=0 Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.901588 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" event={"ID":"34f0f91e-2de6-4e13-974c-917d871c3a43","Type":"ContainerDied","Data":"8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.901617 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" event={"ID":"34f0f91e-2de6-4e13-974c-917d871c3a43","Type":"ContainerStarted","Data":"df3b45024336a5d8fa0fafae1bcbb491dac2ea23d6283d3d07dc67b388091b61"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.902934 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"9ed6ca1ba0386f035932c048a7d8c2f5fc8271e5baa5c2f4de8bd39a83bd3c60"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.906333 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.906363 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.906386 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"b75d2314dc048f051c0af95cd09d5f9da14b9f9d02b81d88167cf5ec8e9647c9"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.909077 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.909612 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.909643 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.909658 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"176783ac6fbfceb6c309e97b1cfd5a47e1fff2fae18e0a99e4e18c43b7a71b1d"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.911761 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3" exitCode=0 Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.911900 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.911987 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"3c022f138769ab1d64f295dda0b5d503db70d70899c44ff8fa396d4de1d9824d"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.917857 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.920391 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.920661 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.921866 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5rllg" event={"ID":"1f5d3d8c-cdec-4ed3-bb53-0a5012751616","Type":"ContainerStarted","Data":"a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.921923 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-5rllg" event={"ID":"1f5d3d8c-cdec-4ed3-bb53-0a5012751616","Type":"ContainerStarted","Data":"056a33f9bf87acc5c2a9da8d6463934bd209abe83d1fce4ca8500b7ce33b216b"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.924617 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.924674 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"b1a39a5e3f1d01bf4666f90d90aec8fe54917e2aa9906a70294905b36cb7403e"} Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.925611 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.941136 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.955214 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.975756 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:23 crc kubenswrapper[4787]: I1203 17:13:23.994271 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.016584 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.035595 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.051619 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.070968 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.088656 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.105381 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.120452 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.134974 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.149243 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.162079 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.180786 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.199790 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.215776 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.360300 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.360470 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.360487 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:13:26.360461546 +0000 UTC m=+23.177932505 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.360591 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.360598 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.360634 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.360681 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:26.360672672 +0000 UTC m=+23.178143631 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.360737 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:26.360688222 +0000 UTC m=+23.178159191 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.371094 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-xnl29"] Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.371432 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xnl29" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.373807 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.373937 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.374236 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.374364 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.402855 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.417770 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.437301 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.451881 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.461695 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c61ffb2-8f34-4a8b-9e65-89da492c3382-host\") pod \"node-ca-xnl29\" (UID: \"9c61ffb2-8f34-4a8b-9e65-89da492c3382\") " pod="openshift-image-registry/node-ca-xnl29" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.461743 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9c61ffb2-8f34-4a8b-9e65-89da492c3382-serviceca\") pod \"node-ca-xnl29\" (UID: \"9c61ffb2-8f34-4a8b-9e65-89da492c3382\") " pod="openshift-image-registry/node-ca-xnl29" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.461771 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j5xbx\" (UniqueName: \"kubernetes.io/projected/9c61ffb2-8f34-4a8b-9e65-89da492c3382-kube-api-access-j5xbx\") pod \"node-ca-xnl29\" (UID: \"9c61ffb2-8f34-4a8b-9e65-89da492c3382\") " pod="openshift-image-registry/node-ca-xnl29" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.461790 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.461812 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.461919 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.461937 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.461949 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.461977 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.462011 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.462036 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.461991 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:26.461979417 +0000 UTC m=+23.279450376 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.462088 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:26.462072679 +0000 UTC m=+23.279543638 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.467008 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.480790 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.499773 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.512220 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.526417 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.541062 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.560556 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.562229 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c61ffb2-8f34-4a8b-9e65-89da492c3382-host\") pod \"node-ca-xnl29\" (UID: \"9c61ffb2-8f34-4a8b-9e65-89da492c3382\") " pod="openshift-image-registry/node-ca-xnl29" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.562261 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9c61ffb2-8f34-4a8b-9e65-89da492c3382-serviceca\") pod \"node-ca-xnl29\" (UID: \"9c61ffb2-8f34-4a8b-9e65-89da492c3382\") " pod="openshift-image-registry/node-ca-xnl29" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.562287 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j5xbx\" (UniqueName: \"kubernetes.io/projected/9c61ffb2-8f34-4a8b-9e65-89da492c3382-kube-api-access-j5xbx\") pod \"node-ca-xnl29\" (UID: \"9c61ffb2-8f34-4a8b-9e65-89da492c3382\") " pod="openshift-image-registry/node-ca-xnl29" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.562348 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9c61ffb2-8f34-4a8b-9e65-89da492c3382-host\") pod \"node-ca-xnl29\" (UID: \"9c61ffb2-8f34-4a8b-9e65-89da492c3382\") " pod="openshift-image-registry/node-ca-xnl29" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.563357 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/9c61ffb2-8f34-4a8b-9e65-89da492c3382-serviceca\") pod \"node-ca-xnl29\" (UID: \"9c61ffb2-8f34-4a8b-9e65-89da492c3382\") " pod="openshift-image-registry/node-ca-xnl29" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.571725 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.586513 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j5xbx\" (UniqueName: \"kubernetes.io/projected/9c61ffb2-8f34-4a8b-9e65-89da492c3382-kube-api-access-j5xbx\") pod \"node-ca-xnl29\" (UID: \"9c61ffb2-8f34-4a8b-9e65-89da492c3382\") " pod="openshift-image-registry/node-ca-xnl29" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.599133 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.617587 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.766428 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.767097 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.766507 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.767406 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.766476 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:24 crc kubenswrapper[4787]: E1203 17:13:24.767674 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.812416 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-xnl29" Dec 03 17:13:24 crc kubenswrapper[4787]: W1203 17:13:24.834723 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c61ffb2_8f34_4a8b_9e65_89da492c3382.slice/crio-123dc23a0158cbff23e17de0b210804507cc50e3a6fa31d453d6e70e9289bcf8 WatchSource:0}: Error finding container 123dc23a0158cbff23e17de0b210804507cc50e3a6fa31d453d6e70e9289bcf8: Status 404 returned error can't find the container with id 123dc23a0158cbff23e17de0b210804507cc50e3a6fa31d453d6e70e9289bcf8 Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.912282 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.937862 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3"} Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.937917 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4"} Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.937929 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e"} Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.941304 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" event={"ID":"34f0f91e-2de6-4e13-974c-917d871c3a43","Type":"ContainerStarted","Data":"ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820"} Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.946474 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xnl29" event={"ID":"9c61ffb2-8f34-4a8b-9e65-89da492c3382","Type":"ContainerStarted","Data":"123dc23a0158cbff23e17de0b210804507cc50e3a6fa31d453d6e70e9289bcf8"} Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.949858 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.951277 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.955230 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.971996 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.984894 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:24 crc kubenswrapper[4787]: I1203 17:13:24.999143 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.020721 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.032958 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.047254 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.063083 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.075636 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.087069 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.090277 4787 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.092198 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.092238 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.092250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.092409 4787 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.099662 4787 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.099965 4787 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.101219 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.101288 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.101300 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.101322 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.101336 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.105422 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: E1203 17:13:25.119185 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.122399 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.122432 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.122444 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.122466 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.122480 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: E1203 17:13:25.136642 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.140143 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.140180 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.140189 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.140204 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.140218 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.147881 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: E1203 17:13:25.150627 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.154943 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.154971 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.154981 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.155026 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.155042 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: E1203 17:13:25.170783 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.177998 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.178048 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.178060 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.178077 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.178089 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.187370 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: E1203 17:13:25.190957 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: E1203 17:13:25.191106 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.192424 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.192455 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.192466 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.192482 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.192494 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.226864 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.267799 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.295395 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.295463 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.295476 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.295501 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.295515 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.305683 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.349812 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.384303 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.398977 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.399077 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.399091 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.399115 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.399134 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.427996 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.469232 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.502968 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.503046 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.503059 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.503081 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.503095 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.524862 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.549856 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.590368 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.605836 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.605902 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.605914 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.605935 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.605950 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.634554 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.669373 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.709835 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.709857 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.709904 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.710007 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.710059 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.710080 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.753332 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.785462 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.813469 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.813544 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.813562 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.813586 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.813601 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.829494 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.916368 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.916421 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.916435 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.916454 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.916467 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:25Z","lastTransitionTime":"2025-12-03T17:13:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.968262 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-xnl29" event={"ID":"9c61ffb2-8f34-4a8b-9e65-89da492c3382","Type":"ContainerStarted","Data":"23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.974985 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.975056 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.975071 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.977678 4787 generic.go:334] "Generic (PLEG): container finished" podID="34f0f91e-2de6-4e13-974c-917d871c3a43" containerID="ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820" exitCode=0 Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.977736 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" event={"ID":"34f0f91e-2de6-4e13-974c-917d871c3a43","Type":"ContainerDied","Data":"ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820"} Dec 03 17:13:25 crc kubenswrapper[4787]: I1203 17:13:25.991255 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:25Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.016937 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.020202 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.020261 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.020280 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.020306 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.020364 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:26Z","lastTransitionTime":"2025-12-03T17:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.046214 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.059588 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.073926 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.087605 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.113116 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.124286 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.124326 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.124338 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.124354 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.124365 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:26Z","lastTransitionTime":"2025-12-03T17:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.149418 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.188507 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.226479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.226521 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.226531 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.226548 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.226558 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:26Z","lastTransitionTime":"2025-12-03T17:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.227830 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.267289 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.310533 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.329383 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.329426 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.329435 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.329448 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.329458 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:26Z","lastTransitionTime":"2025-12-03T17:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.347567 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.380869 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.380974 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.381046 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.381112 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.381136 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:13:30.381099554 +0000 UTC m=+27.198570543 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.381186 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:30.381169786 +0000 UTC m=+27.198640885 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.381272 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.381364 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:30.38134664 +0000 UTC m=+27.198817599 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.384563 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.430204 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.431539 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.431588 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.431602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.431625 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.431648 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:26Z","lastTransitionTime":"2025-12-03T17:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.465746 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.482781 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.482878 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.483051 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.483071 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.483090 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.483097 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.483105 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.483110 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.483181 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:30.483159769 +0000 UTC m=+27.300630738 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.483202 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:30.48319393 +0000 UTC m=+27.300664899 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.506823 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.535002 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.535084 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.535097 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.535116 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.535129 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:26Z","lastTransitionTime":"2025-12-03T17:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.545588 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.586373 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.638617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.638714 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.638733 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.638763 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.638785 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:26Z","lastTransitionTime":"2025-12-03T17:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.646165 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.675835 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.710513 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.741664 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.741727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.741751 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.741783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.741837 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:26Z","lastTransitionTime":"2025-12-03T17:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.747937 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.765938 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.765944 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.766233 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.765947 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.766420 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:26 crc kubenswrapper[4787]: E1203 17:13:26.766509 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.792838 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.831177 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.846390 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.846491 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.846507 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.846534 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.846549 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:26Z","lastTransitionTime":"2025-12-03T17:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.870836 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.920360 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.946910 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.948925 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.948969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.948985 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.949004 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.949034 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:26Z","lastTransitionTime":"2025-12-03T17:13:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.985722 4787 generic.go:334] "Generic (PLEG): container finished" podID="34f0f91e-2de6-4e13-974c-917d871c3a43" containerID="afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8" exitCode=0 Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.985899 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" event={"ID":"34f0f91e-2de6-4e13-974c-917d871c3a43","Type":"ContainerDied","Data":"afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8"} Dec 03 17:13:26 crc kubenswrapper[4787]: I1203 17:13:26.990893 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.026780 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.051516 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.051564 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.051577 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.051594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.051604 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:27Z","lastTransitionTime":"2025-12-03T17:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.070134 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.112953 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.146823 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.156436 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.156506 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.156530 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.156564 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.156587 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:27Z","lastTransitionTime":"2025-12-03T17:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.199645 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.231032 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.259159 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.259200 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.259209 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.259225 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.259242 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:27Z","lastTransitionTime":"2025-12-03T17:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.270162 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.306071 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.353824 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.362498 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.362559 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.362571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.362597 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.362611 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:27Z","lastTransitionTime":"2025-12-03T17:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.387700 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.432298 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.465730 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.465778 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.465794 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.465816 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.465827 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:27Z","lastTransitionTime":"2025-12-03T17:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.473828 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.506861 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.551972 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.567857 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.567896 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.567909 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.567924 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.567934 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:27Z","lastTransitionTime":"2025-12-03T17:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.589508 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.644769 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.671321 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.671380 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.671396 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.671415 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.671429 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:27Z","lastTransitionTime":"2025-12-03T17:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.774628 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.775063 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.775074 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.775094 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.775105 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:27Z","lastTransitionTime":"2025-12-03T17:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.877994 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.878050 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.878062 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.878079 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.878092 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:27Z","lastTransitionTime":"2025-12-03T17:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.981225 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.981279 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.981293 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.981315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.981331 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:27Z","lastTransitionTime":"2025-12-03T17:13:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.993899 4787 generic.go:334] "Generic (PLEG): container finished" podID="34f0f91e-2de6-4e13-974c-917d871c3a43" containerID="14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5" exitCode=0 Dec 03 17:13:27 crc kubenswrapper[4787]: I1203 17:13:27.993993 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" event={"ID":"34f0f91e-2de6-4e13-974c-917d871c3a43","Type":"ContainerDied","Data":"14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5"} Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.000899 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c"} Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.005320 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b"} Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.038118 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.054499 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.084187 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.084261 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.084282 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.084307 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.084330 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:28Z","lastTransitionTime":"2025-12-03T17:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.087562 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.111315 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.135450 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.163675 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.180422 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.195908 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.195963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.195983 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.196009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.196058 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:28Z","lastTransitionTime":"2025-12-03T17:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.200976 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.268376 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.291680 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.300348 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.300386 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.300396 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.300413 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.300428 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:28Z","lastTransitionTime":"2025-12-03T17:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.305888 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.325103 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.335869 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.351364 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.365461 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.387762 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.402487 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.402537 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.402549 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.402566 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.402577 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:28Z","lastTransitionTime":"2025-12-03T17:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.402820 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.423261 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.438655 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.453113 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.472606 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.506225 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.506346 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.506359 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.506377 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.506389 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:28Z","lastTransitionTime":"2025-12-03T17:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.509701 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.550850 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.588678 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.609865 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.609947 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.609972 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.610004 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.610101 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:28Z","lastTransitionTime":"2025-12-03T17:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.627750 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.674118 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.709884 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.713477 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.713551 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.713575 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.713605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.713629 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:28Z","lastTransitionTime":"2025-12-03T17:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.753763 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.765519 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:28 crc kubenswrapper[4787]: E1203 17:13:28.766191 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.765619 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.765528 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:28 crc kubenswrapper[4787]: E1203 17:13:28.766354 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:28 crc kubenswrapper[4787]: E1203 17:13:28.766467 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.787191 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.816159 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.816190 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.816199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.816213 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.816223 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:28Z","lastTransitionTime":"2025-12-03T17:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.829905 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.918785 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.918847 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.918864 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.918888 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:28 crc kubenswrapper[4787]: I1203 17:13:28.918906 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:28Z","lastTransitionTime":"2025-12-03T17:13:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.013898 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" event={"ID":"34f0f91e-2de6-4e13-974c-917d871c3a43","Type":"ContainerStarted","Data":"8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444"} Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.021394 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.021462 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.021484 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.021514 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.021535 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:29Z","lastTransitionTime":"2025-12-03T17:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.045318 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.061366 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.080905 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.095047 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.114463 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.123854 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.123911 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.123931 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.123957 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.123979 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:29Z","lastTransitionTime":"2025-12-03T17:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.137403 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.161298 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.181730 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.200885 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.227884 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.228355 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.228540 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.228870 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.229087 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:29Z","lastTransitionTime":"2025-12-03T17:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.233702 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.283554 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.313077 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.332779 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.332821 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.332829 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.332844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.332853 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:29Z","lastTransitionTime":"2025-12-03T17:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.348298 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.389898 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.432571 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:29Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.436353 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.436431 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.436444 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.436464 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.436480 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:29Z","lastTransitionTime":"2025-12-03T17:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.538578 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.538635 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.538647 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.538671 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.538691 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:29Z","lastTransitionTime":"2025-12-03T17:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.642834 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.642891 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.642906 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.642931 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.642947 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:29Z","lastTransitionTime":"2025-12-03T17:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.745799 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.745848 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.745862 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.745882 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.745893 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:29Z","lastTransitionTime":"2025-12-03T17:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.848737 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.848771 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.848781 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.848795 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.848808 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:29Z","lastTransitionTime":"2025-12-03T17:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.951470 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.951531 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.951547 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.951568 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:29 crc kubenswrapper[4787]: I1203 17:13:29.951580 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:29Z","lastTransitionTime":"2025-12-03T17:13:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.021054 4787 generic.go:334] "Generic (PLEG): container finished" podID="34f0f91e-2de6-4e13-974c-917d871c3a43" containerID="8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444" exitCode=0 Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.021137 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" event={"ID":"34f0f91e-2de6-4e13-974c-917d871c3a43","Type":"ContainerDied","Data":"8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.026351 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.026675 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.026708 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.040656 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.047397 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.048271 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.054699 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.054750 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.054761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.054806 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.054822 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:30Z","lastTransitionTime":"2025-12-03T17:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.056646 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.070101 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.087392 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.102237 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.118883 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.139306 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.150302 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.158844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.158915 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.158935 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.158959 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.158972 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:30Z","lastTransitionTime":"2025-12-03T17:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.169121 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.181847 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.194793 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.208241 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.221787 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.237611 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.258081 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.262664 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.262695 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.262708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.262731 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.262744 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:30Z","lastTransitionTime":"2025-12-03T17:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.270261 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.284780 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.298165 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.317747 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.335550 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.353297 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.368283 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.368571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.368688 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.368823 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.368948 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:30Z","lastTransitionTime":"2025-12-03T17:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.372808 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.387937 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.407066 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.425258 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.425439 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:13:38.425402837 +0000 UTC m=+35.242873806 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.425519 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.425591 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.425720 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.425813 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:38.425791227 +0000 UTC m=+35.243262196 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.425723 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.425898 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:38.42588719 +0000 UTC m=+35.243358159 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.427167 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.468178 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.473526 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.473594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.473616 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.473644 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.473664 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:30Z","lastTransitionTime":"2025-12-03T17:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.506634 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.526785 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.526839 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.527038 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.527070 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.527089 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.527161 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:38.527142644 +0000 UTC m=+35.344613603 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.527156 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.527216 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.527238 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.527332 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:38.527304558 +0000 UTC m=+35.344775547 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.553714 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.576617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.576679 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.576696 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.576722 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.576741 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:30Z","lastTransitionTime":"2025-12-03T17:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.587138 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.627495 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:30Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.679603 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.679645 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.679655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.679672 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.679684 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:30Z","lastTransitionTime":"2025-12-03T17:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.766063 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.766187 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.766277 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.766523 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.766656 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:30 crc kubenswrapper[4787]: E1203 17:13:30.766857 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.783658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.783762 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.783786 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.783821 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.783848 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:30Z","lastTransitionTime":"2025-12-03T17:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.886929 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.886979 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.886988 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.887002 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.887012 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:30Z","lastTransitionTime":"2025-12-03T17:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.991962 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.991998 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.992006 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.992034 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:30 crc kubenswrapper[4787]: I1203 17:13:30.992043 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:30Z","lastTransitionTime":"2025-12-03T17:13:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.032780 4787 generic.go:334] "Generic (PLEG): container finished" podID="34f0f91e-2de6-4e13-974c-917d871c3a43" containerID="bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3" exitCode=0 Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.032866 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" event={"ID":"34f0f91e-2de6-4e13-974c-917d871c3a43","Type":"ContainerDied","Data":"bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3"} Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.032885 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.054994 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.068798 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.085364 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.094884 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.094922 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.094934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.094951 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.094962 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:31Z","lastTransitionTime":"2025-12-03T17:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.098317 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.113761 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.126658 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.149457 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.161930 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.179165 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.193354 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.201715 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.201790 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.201813 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.201838 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.201857 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:31Z","lastTransitionTime":"2025-12-03T17:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.211492 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.236425 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.252271 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.268100 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.285740 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:31Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.304489 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.304524 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.304533 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.304546 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.304555 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:31Z","lastTransitionTime":"2025-12-03T17:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.406764 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.406816 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.406827 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.406844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.406853 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:31Z","lastTransitionTime":"2025-12-03T17:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.509731 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.509800 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.509819 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.509844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.509861 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:31Z","lastTransitionTime":"2025-12-03T17:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.613072 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.613138 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.613162 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.613188 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.613206 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:31Z","lastTransitionTime":"2025-12-03T17:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.715803 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.715888 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.715917 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.715951 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.715978 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:31Z","lastTransitionTime":"2025-12-03T17:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.818371 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.818426 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.818443 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.818465 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.818489 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:31Z","lastTransitionTime":"2025-12-03T17:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.921920 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.922013 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.922078 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.922107 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:31 crc kubenswrapper[4787]: I1203 17:13:31.922131 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:31Z","lastTransitionTime":"2025-12-03T17:13:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.025476 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.025536 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.025553 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.025576 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.025593 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:32Z","lastTransitionTime":"2025-12-03T17:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.042903 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" event={"ID":"34f0f91e-2de6-4e13-974c-917d871c3a43","Type":"ContainerStarted","Data":"3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09"} Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.043010 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.066171 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.086311 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.120143 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.128494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.128572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.128590 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.128615 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.128631 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:32Z","lastTransitionTime":"2025-12-03T17:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.141793 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.165007 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.183733 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.199935 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.217978 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.231287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.231340 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.231366 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.231386 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.231399 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:32Z","lastTransitionTime":"2025-12-03T17:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.236709 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.247886 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.260240 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.272704 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.283131 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.298708 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.306871 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:32Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.336546 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.336586 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.336594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.336607 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.336617 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:32Z","lastTransitionTime":"2025-12-03T17:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.438944 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.438999 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.439011 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.439056 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.439073 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:32Z","lastTransitionTime":"2025-12-03T17:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.541144 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.541178 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.541190 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.541205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.541216 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:32Z","lastTransitionTime":"2025-12-03T17:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.643715 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.643754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.643766 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.643783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.643794 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:32Z","lastTransitionTime":"2025-12-03T17:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.746402 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.746432 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.746445 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.746461 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.746473 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:32Z","lastTransitionTime":"2025-12-03T17:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.765369 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:32 crc kubenswrapper[4787]: E1203 17:13:32.765500 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.765372 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:32 crc kubenswrapper[4787]: E1203 17:13:32.765633 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.765845 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:32 crc kubenswrapper[4787]: E1203 17:13:32.765913 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.848936 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.848991 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.849006 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.849053 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.849073 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:32Z","lastTransitionTime":"2025-12-03T17:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.952224 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.952635 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.952647 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.952668 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:32 crc kubenswrapper[4787]: I1203 17:13:32.952683 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:32Z","lastTransitionTime":"2025-12-03T17:13:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.055366 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.055423 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.055437 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.055458 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.055475 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:33Z","lastTransitionTime":"2025-12-03T17:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.159245 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.159293 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.159305 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.159324 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.159335 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:33Z","lastTransitionTime":"2025-12-03T17:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.262545 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.262594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.262609 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.262633 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.262650 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:33Z","lastTransitionTime":"2025-12-03T17:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.365881 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.365920 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.365930 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.365945 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.365955 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:33Z","lastTransitionTime":"2025-12-03T17:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.468504 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.468580 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.468600 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.468637 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.468662 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:33Z","lastTransitionTime":"2025-12-03T17:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.572344 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.572395 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.572405 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.572433 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.572446 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:33Z","lastTransitionTime":"2025-12-03T17:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.675608 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.675655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.675668 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.675683 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.675695 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:33Z","lastTransitionTime":"2025-12-03T17:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.778055 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.778132 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.778156 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.778186 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.778207 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:33Z","lastTransitionTime":"2025-12-03T17:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.810415 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.842079 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.855944 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.867640 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.881334 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.881364 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.881372 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.881388 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.881399 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:33Z","lastTransitionTime":"2025-12-03T17:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.893175 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.907908 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.924824 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.938209 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.951003 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.967122 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.979552 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.983630 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.983692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.983703 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.983716 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.983726 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:33Z","lastTransitionTime":"2025-12-03T17:13:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:33 crc kubenswrapper[4787]: I1203 17:13:33.998549 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.013203 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.028927 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.042146 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.052153 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/0.log" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.055177 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580" exitCode=1 Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.055221 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580"} Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.055947 4787 scope.go:117] "RemoveContainer" containerID="b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.074883 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.087115 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.087157 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.087167 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.087182 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.087192 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:34Z","lastTransitionTime":"2025-12-03T17:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.091267 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.110485 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.130130 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.148718 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.164405 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.179584 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.189655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.189725 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.189746 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.189775 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.189796 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:34Z","lastTransitionTime":"2025-12-03T17:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.193715 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.215122 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:33Z\\\",\\\"message\\\":\\\"dler 5 for removal\\\\nI1203 17:13:33.094104 6073 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.094122 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 17:13:33.093989 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 17:13:33.094057 6073 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094269 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 17:13:33.094270 6073 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.093998 6073 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094112 6073 factory.go:656] Stopping watch factory\\\\nI1203 17:13:33.094717 6073 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 17:13:33.095182 6073 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.229752 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.246497 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.262432 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.288505 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.293921 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.294284 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.294622 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.294889 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.295109 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:34Z","lastTransitionTime":"2025-12-03T17:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.303603 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.325505 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.399317 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.399360 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.399374 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.399395 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.399412 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:34Z","lastTransitionTime":"2025-12-03T17:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.502814 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.503241 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.503440 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.503608 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.503765 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:34Z","lastTransitionTime":"2025-12-03T17:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.606759 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.607123 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.607300 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.607543 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.607714 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:34Z","lastTransitionTime":"2025-12-03T17:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.710907 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.710993 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.711022 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.711097 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.711122 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:34Z","lastTransitionTime":"2025-12-03T17:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.765476 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.765827 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.765465 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:34 crc kubenswrapper[4787]: E1203 17:13:34.765858 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:34 crc kubenswrapper[4787]: E1203 17:13:34.766085 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:34 crc kubenswrapper[4787]: E1203 17:13:34.766249 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.814686 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.814729 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.814741 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.814757 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.814769 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:34Z","lastTransitionTime":"2025-12-03T17:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.917414 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.917461 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.917472 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.917492 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.917505 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:34Z","lastTransitionTime":"2025-12-03T17:13:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.943836 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7"] Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.944777 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.947103 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.947925 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.966200 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:34 crc kubenswrapper[4787]: I1203 17:13:34.980341 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.002370 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:33Z\\\",\\\"message\\\":\\\"dler 5 for removal\\\\nI1203 17:13:33.094104 6073 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.094122 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 17:13:33.093989 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 17:13:33.094057 6073 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094269 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 17:13:33.094270 6073 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.093998 6073 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094112 6073 factory.go:656] Stopping watch factory\\\\nI1203 17:13:33.094717 6073 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 17:13:33.095182 6073 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.012445 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.020620 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.020674 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.020691 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.020711 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.020725 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.025863 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.037669 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.048749 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.059489 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/0.log" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.061556 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.061659 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.067427 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.073435 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x886\" (UniqueName: \"kubernetes.io/projected/ef833e74-df41-4fe8-bec4-b5cf3541aa99-kube-api-access-2x886\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.073589 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef833e74-df41-4fe8-bec4-b5cf3541aa99-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.073737 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef833e74-df41-4fe8-bec4-b5cf3541aa99-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.073867 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef833e74-df41-4fe8-bec4-b5cf3541aa99-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.081637 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.095380 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.105176 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.119477 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.123199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.123243 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.123254 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.123270 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.123281 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.136506 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.149877 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.161744 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.174576 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef833e74-df41-4fe8-bec4-b5cf3541aa99-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.174643 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef833e74-df41-4fe8-bec4-b5cf3541aa99-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.175476 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x886\" (UniqueName: \"kubernetes.io/projected/ef833e74-df41-4fe8-bec4-b5cf3541aa99-kube-api-access-2x886\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.175536 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef833e74-df41-4fe8-bec4-b5cf3541aa99-env-overrides\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.175556 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef833e74-df41-4fe8-bec4-b5cf3541aa99-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.176213 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/ef833e74-df41-4fe8-bec4-b5cf3541aa99-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.177104 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.183646 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/ef833e74-df41-4fe8-bec4-b5cf3541aa99-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.198482 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x886\" (UniqueName: \"kubernetes.io/projected/ef833e74-df41-4fe8-bec4-b5cf3541aa99-kube-api-access-2x886\") pod \"ovnkube-control-plane-749d76644c-6zkg7\" (UID: \"ef833e74-df41-4fe8-bec4-b5cf3541aa99\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.212085 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.225067 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.225125 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.225292 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.225317 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.225341 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.225354 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.242503 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.253009 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.265151 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.268249 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: W1203 17:13:35.279038 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef833e74_df41_4fe8_bec4_b5cf3541aa99.slice/crio-3bdc0c6d0284d3413fb30c94046a25c1a669bbe0727b9a68208d478b0d103c3f WatchSource:0}: Error finding container 3bdc0c6d0284d3413fb30c94046a25c1a669bbe0727b9a68208d478b0d103c3f: Status 404 returned error can't find the container with id 3bdc0c6d0284d3413fb30c94046a25c1a669bbe0727b9a68208d478b0d103c3f Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.284973 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.302006 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.317861 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.328602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.328631 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.328642 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.328658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.328670 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.333678 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.350880 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.370176 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.376017 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.376067 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.376078 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.376095 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.376106 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: E1203 17:13:35.390381 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.391238 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:33Z\\\",\\\"message\\\":\\\"dler 5 for removal\\\\nI1203 17:13:33.094104 6073 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.094122 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 17:13:33.093989 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 17:13:33.094057 6073 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094269 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 17:13:33.094270 6073 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.093998 6073 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094112 6073 factory.go:656] Stopping watch factory\\\\nI1203 17:13:33.094717 6073 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 17:13:33.095182 6073 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.396304 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.396364 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.396377 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.396396 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.396408 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.403229 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: E1203 17:13:35.408464 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.412313 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.412370 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.412418 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.412439 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.412453 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.420218 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: E1203 17:13:35.425655 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.429106 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.429143 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.429156 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.429173 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.429187 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.432372 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: E1203 17:13:35.441592 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.446013 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.447555 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.447599 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.447615 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.447636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.447649 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: E1203 17:13:35.461578 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:35Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:35Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:35 crc kubenswrapper[4787]: E1203 17:13:35.461716 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.463697 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.463732 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.463742 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.463759 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.463772 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.566809 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.566850 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.566859 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.566874 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.566883 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.669772 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.669830 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.669845 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.669866 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.669878 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.772482 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.772556 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.772571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.772590 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.772605 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.881482 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.881527 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.881536 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.881553 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.881565 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.983553 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.983596 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.983605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.983619 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:35 crc kubenswrapper[4787]: I1203 17:13:35.983630 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:35Z","lastTransitionTime":"2025-12-03T17:13:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.066690 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" event={"ID":"ef833e74-df41-4fe8-bec4-b5cf3541aa99","Type":"ContainerStarted","Data":"3bdc0c6d0284d3413fb30c94046a25c1a669bbe0727b9a68208d478b0d103c3f"} Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.085658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.085740 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.085772 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.085805 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.085827 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:36Z","lastTransitionTime":"2025-12-03T17:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.189070 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.189141 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.189163 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.189192 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.189215 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:36Z","lastTransitionTime":"2025-12-03T17:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.292314 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.292360 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.292371 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.292388 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.292399 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:36Z","lastTransitionTime":"2025-12-03T17:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.395711 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.395766 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.395778 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.395796 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.395809 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:36Z","lastTransitionTime":"2025-12-03T17:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.426962 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-m9tr6"] Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.427482 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:36 crc kubenswrapper[4787]: E1203 17:13:36.427615 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.442641 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.460648 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.472636 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.486621 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.486672 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbgkj\" (UniqueName: \"kubernetes.io/projected/50c2569b-2a14-4112-82e4-afc683aa36a7-kube-api-access-wbgkj\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.494171 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.497557 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.497604 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.497617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.497641 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.497680 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:36Z","lastTransitionTime":"2025-12-03T17:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.508974 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.523929 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.536658 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.562494 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.574676 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.587293 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:36 crc kubenswrapper[4787]: E1203 17:13:36.587452 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:36 crc kubenswrapper[4787]: E1203 17:13:36.587633 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs podName:50c2569b-2a14-4112-82e4-afc683aa36a7 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:37.087610178 +0000 UTC m=+33.905081147 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs") pod "network-metrics-daemon-m9tr6" (UID: "50c2569b-2a14-4112-82e4-afc683aa36a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.587531 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbgkj\" (UniqueName: \"kubernetes.io/projected/50c2569b-2a14-4112-82e4-afc683aa36a7-kube-api-access-wbgkj\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.592434 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.600611 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.600891 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.600969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.601062 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.601274 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:36Z","lastTransitionTime":"2025-12-03T17:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.605097 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.614867 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbgkj\" (UniqueName: \"kubernetes.io/projected/50c2569b-2a14-4112-82e4-afc683aa36a7-kube-api-access-wbgkj\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.621329 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.637588 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.644865 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.654072 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.669895 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.694806 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:33Z\\\",\\\"message\\\":\\\"dler 5 for removal\\\\nI1203 17:13:33.094104 6073 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.094122 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 17:13:33.093989 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 17:13:33.094057 6073 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094269 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 17:13:33.094270 6073 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.093998 6073 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094112 6073 factory.go:656] Stopping watch factory\\\\nI1203 17:13:33.094717 6073 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 17:13:33.095182 6073 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.706514 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.707107 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.707276 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.707418 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.707554 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.707670 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:36Z","lastTransitionTime":"2025-12-03T17:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.717337 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.748391 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:33Z\\\",\\\"message\\\":\\\"dler 5 for removal\\\\nI1203 17:13:33.094104 6073 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.094122 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 17:13:33.093989 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 17:13:33.094057 6073 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094269 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 17:13:33.094270 6073 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.093998 6073 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094112 6073 factory.go:656] Stopping watch factory\\\\nI1203 17:13:33.094717 6073 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 17:13:33.095182 6073 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.765579 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.765589 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: E1203 17:13:36.765696 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.765579 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:36 crc kubenswrapper[4787]: E1203 17:13:36.765841 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.765850 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:36 crc kubenswrapper[4787]: E1203 17:13:36.766095 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.788941 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.807971 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.810173 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.810215 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.810228 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.810248 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.810261 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:36Z","lastTransitionTime":"2025-12-03T17:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.827372 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.839654 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.870751 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.889892 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.902762 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.912416 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.912452 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.912460 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.912475 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.912485 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:36Z","lastTransitionTime":"2025-12-03T17:13:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.916769 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.929700 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.944109 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.955860 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.969818 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.979714 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:36 crc kubenswrapper[4787]: I1203 17:13:36.991741 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:36Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.014240 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.014275 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.014283 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.014294 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.014304 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:37Z","lastTransitionTime":"2025-12-03T17:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.072384 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" event={"ID":"ef833e74-df41-4fe8-bec4-b5cf3541aa99","Type":"ContainerStarted","Data":"d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.072442 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" event={"ID":"ef833e74-df41-4fe8-bec4-b5cf3541aa99","Type":"ContainerStarted","Data":"f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.075313 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/1.log" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.075928 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/0.log" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.079283 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679" exitCode=1 Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.079354 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.079428 4787 scope.go:117] "RemoveContainer" containerID="b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.080722 4787 scope.go:117] "RemoveContainer" containerID="0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679" Dec 03 17:13:37 crc kubenswrapper[4787]: E1203 17:13:37.081007 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.090691 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.091532 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:37 crc kubenswrapper[4787]: E1203 17:13:37.091738 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:37 crc kubenswrapper[4787]: E1203 17:13:37.091827 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs podName:50c2569b-2a14-4112-82e4-afc683aa36a7 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:38.091804491 +0000 UTC m=+34.909275450 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs") pod "network-metrics-daemon-m9tr6" (UID: "50c2569b-2a14-4112-82e4-afc683aa36a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.104457 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.115923 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.116937 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.116990 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.117009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.117064 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.117085 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:37Z","lastTransitionTime":"2025-12-03T17:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.140835 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.161281 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.174756 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.189842 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.199850 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.213236 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.219867 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.219925 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.219939 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.219957 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.219968 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:37Z","lastTransitionTime":"2025-12-03T17:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.226483 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.237152 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.252427 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.263362 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.275442 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.285000 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.301391 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:33Z\\\",\\\"message\\\":\\\"dler 5 for removal\\\\nI1203 17:13:33.094104 6073 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.094122 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 17:13:33.093989 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 17:13:33.094057 6073 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094269 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 17:13:33.094270 6073 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.093998 6073 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094112 6073 factory.go:656] Stopping watch factory\\\\nI1203 17:13:33.094717 6073 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 17:13:33.095182 6073 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.310134 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.318321 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.322309 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.322371 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.322393 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.322420 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.322437 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:37Z","lastTransitionTime":"2025-12-03T17:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.330961 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.342264 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.358677 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:33Z\\\",\\\"message\\\":\\\"dler 5 for removal\\\\nI1203 17:13:33.094104 6073 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.094122 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 17:13:33.093989 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 17:13:33.094057 6073 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094269 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 17:13:33.094270 6073 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.093998 6073 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094112 6073 factory.go:656] Stopping watch factory\\\\nI1203 17:13:33.094717 6073 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 17:13:33.095182 6073 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\" fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879315 6249 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 17:13:35.879331 6249 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 17:13:35.879422 6249 factory.go:656] Stopping watch factory\\\\nI1203 17:13:35.879438 6249 ovnkube.go:599] Stopped ovnkube\\\\nI1203 17:13:35.879398 6249 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879469 6249 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 17:13:35.879479 6249 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:13:35.879572 6249 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.368740 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.378583 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.389489 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.401499 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.424336 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.425289 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.425343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.425356 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.425378 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.425390 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:37Z","lastTransitionTime":"2025-12-03T17:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.437383 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.450681 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.465508 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.487629 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.505030 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.522994 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.528247 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.528292 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.528302 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.528318 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.528329 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:37Z","lastTransitionTime":"2025-12-03T17:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.544506 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.568940 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:37Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.630562 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.630637 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.630661 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.630694 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.630718 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:37Z","lastTransitionTime":"2025-12-03T17:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.733367 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.733423 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.733441 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.733461 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.733475 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:37Z","lastTransitionTime":"2025-12-03T17:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.765891 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:37 crc kubenswrapper[4787]: E1203 17:13:37.766086 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.836977 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.837080 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.837106 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.837136 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.837161 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:37Z","lastTransitionTime":"2025-12-03T17:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.940251 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.940320 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.940344 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.940373 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:37 crc kubenswrapper[4787]: I1203 17:13:37.940396 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:37Z","lastTransitionTime":"2025-12-03T17:13:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.042895 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.042948 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.042965 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.042990 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.043009 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:38Z","lastTransitionTime":"2025-12-03T17:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.089444 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/1.log" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.103199 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.103414 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.103529 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs podName:50c2569b-2a14-4112-82e4-afc683aa36a7 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:40.103493166 +0000 UTC m=+36.920964165 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs") pod "network-metrics-daemon-m9tr6" (UID: "50c2569b-2a14-4112-82e4-afc683aa36a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.146099 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.146172 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.146195 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.146227 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.146249 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:38Z","lastTransitionTime":"2025-12-03T17:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.249089 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.249143 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.249160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.249186 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.249204 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:38Z","lastTransitionTime":"2025-12-03T17:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.352544 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.352602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.352619 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.352642 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.352659 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:38Z","lastTransitionTime":"2025-12-03T17:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.456472 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.456541 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.456567 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.456597 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.456618 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:38Z","lastTransitionTime":"2025-12-03T17:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.508842 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.509104 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.509191 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.509223 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.509316 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:13:54.509275471 +0000 UTC m=+51.326746470 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.509325 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.509366 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:54.509344093 +0000 UTC m=+51.326815082 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.509409 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:54.509388814 +0000 UTC m=+51.326859803 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.560050 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.560120 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.560157 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.560188 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.560210 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:38Z","lastTransitionTime":"2025-12-03T17:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.610844 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.610936 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.611232 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.611279 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.611307 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.611242 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.611421 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.611445 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.611396 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:54.611365657 +0000 UTC m=+51.428836646 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.611544 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:54.611519171 +0000 UTC m=+51.428990170 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.663790 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.663848 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.663865 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.663889 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.663909 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:38Z","lastTransitionTime":"2025-12-03T17:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.764955 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.764956 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.765100 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.765283 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.765383 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:38 crc kubenswrapper[4787]: E1203 17:13:38.765476 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.767378 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.767452 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.767478 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.767508 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.767534 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:38Z","lastTransitionTime":"2025-12-03T17:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.871074 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.871131 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.871147 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.871163 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.871176 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:38Z","lastTransitionTime":"2025-12-03T17:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.973807 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.973837 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.973846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.973858 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:38 crc kubenswrapper[4787]: I1203 17:13:38.973870 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:38Z","lastTransitionTime":"2025-12-03T17:13:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.076263 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.076333 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.076358 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.076391 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.076417 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:39Z","lastTransitionTime":"2025-12-03T17:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.179636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.179709 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.179732 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.179761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.179780 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:39Z","lastTransitionTime":"2025-12-03T17:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.283619 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.283723 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.283743 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.283768 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.283788 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:39Z","lastTransitionTime":"2025-12-03T17:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.386114 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.386206 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.386228 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.386254 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.386271 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:39Z","lastTransitionTime":"2025-12-03T17:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.489459 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.489565 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.489584 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.489607 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.489625 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:39Z","lastTransitionTime":"2025-12-03T17:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.593203 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.593384 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.593413 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.593448 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.593476 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:39Z","lastTransitionTime":"2025-12-03T17:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.696318 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.696390 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.696411 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.696439 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.696461 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:39Z","lastTransitionTime":"2025-12-03T17:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.766119 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:39 crc kubenswrapper[4787]: E1203 17:13:39.766300 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.798838 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.798895 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.798915 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.798938 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.798961 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:39Z","lastTransitionTime":"2025-12-03T17:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.902842 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.903281 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.903542 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.903768 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:39 crc kubenswrapper[4787]: I1203 17:13:39.903933 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:39Z","lastTransitionTime":"2025-12-03T17:13:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.007116 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.007178 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.007196 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.007219 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.007238 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:40Z","lastTransitionTime":"2025-12-03T17:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.113245 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.113320 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.113343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.113376 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.113398 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:40Z","lastTransitionTime":"2025-12-03T17:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.127907 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:40 crc kubenswrapper[4787]: E1203 17:13:40.128118 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:40 crc kubenswrapper[4787]: E1203 17:13:40.128219 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs podName:50c2569b-2a14-4112-82e4-afc683aa36a7 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:44.12819332 +0000 UTC m=+40.945664319 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs") pod "network-metrics-daemon-m9tr6" (UID: "50c2569b-2a14-4112-82e4-afc683aa36a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.216600 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.216658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.216677 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.216703 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.216720 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:40Z","lastTransitionTime":"2025-12-03T17:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.320062 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.320120 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.320133 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.320149 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.320161 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:40Z","lastTransitionTime":"2025-12-03T17:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.423227 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.423291 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.423314 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.423341 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.423361 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:40Z","lastTransitionTime":"2025-12-03T17:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.526328 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.526436 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.526454 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.526478 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.526494 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:40Z","lastTransitionTime":"2025-12-03T17:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.628791 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.628855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.628875 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.628898 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.628916 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:40Z","lastTransitionTime":"2025-12-03T17:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.731764 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.731837 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.731857 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.731882 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.731903 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:40Z","lastTransitionTime":"2025-12-03T17:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.766113 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.766188 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.766113 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:40 crc kubenswrapper[4787]: E1203 17:13:40.766312 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:40 crc kubenswrapper[4787]: E1203 17:13:40.766460 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:40 crc kubenswrapper[4787]: E1203 17:13:40.766602 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.835251 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.835318 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.835341 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.835369 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.835392 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:40Z","lastTransitionTime":"2025-12-03T17:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.937182 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.937218 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.937227 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.937239 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:40 crc kubenswrapper[4787]: I1203 17:13:40.937250 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:40Z","lastTransitionTime":"2025-12-03T17:13:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.038983 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.039253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.039360 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.039452 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.039528 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:41Z","lastTransitionTime":"2025-12-03T17:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.142044 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.142106 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.142124 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.142151 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.142171 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:41Z","lastTransitionTime":"2025-12-03T17:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.245179 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.245243 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.245268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.245299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.245322 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:41Z","lastTransitionTime":"2025-12-03T17:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.348424 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.348468 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.348480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.348497 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.348508 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:41Z","lastTransitionTime":"2025-12-03T17:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.451717 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.451791 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.451815 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.451844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.451933 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:41Z","lastTransitionTime":"2025-12-03T17:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.554852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.554892 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.554903 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.554920 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.554935 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:41Z","lastTransitionTime":"2025-12-03T17:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.658804 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.658851 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.658863 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.658884 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.658897 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:41Z","lastTransitionTime":"2025-12-03T17:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.761420 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.761492 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.761515 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.761545 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.761567 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:41Z","lastTransitionTime":"2025-12-03T17:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.765896 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:41 crc kubenswrapper[4787]: E1203 17:13:41.766109 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.864193 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.864266 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.864291 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.864325 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.864347 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:41Z","lastTransitionTime":"2025-12-03T17:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.967106 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.967164 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.967181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.967206 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:41 crc kubenswrapper[4787]: I1203 17:13:41.967226 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:41Z","lastTransitionTime":"2025-12-03T17:13:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.070449 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.070535 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.070555 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.070579 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.070597 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:42Z","lastTransitionTime":"2025-12-03T17:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.174598 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.175071 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.175242 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.175356 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.175462 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:42Z","lastTransitionTime":"2025-12-03T17:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.278917 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.278963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.278974 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.278990 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.278999 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:42Z","lastTransitionTime":"2025-12-03T17:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.382080 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.382139 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.382150 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.382176 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.382190 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:42Z","lastTransitionTime":"2025-12-03T17:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.484600 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.484658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.484667 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.484692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.484703 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:42Z","lastTransitionTime":"2025-12-03T17:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.587014 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.587131 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.587161 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.587191 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.587214 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:42Z","lastTransitionTime":"2025-12-03T17:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.689538 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.689599 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.689617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.689640 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.689657 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:42Z","lastTransitionTime":"2025-12-03T17:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.765709 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.765710 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.765898 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:42 crc kubenswrapper[4787]: E1203 17:13:42.765976 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:42 crc kubenswrapper[4787]: E1203 17:13:42.766174 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:42 crc kubenswrapper[4787]: E1203 17:13:42.766290 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.793673 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.793760 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.793783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.793915 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.793933 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:42Z","lastTransitionTime":"2025-12-03T17:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.897394 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.897467 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.897486 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.897509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:42 crc kubenswrapper[4787]: I1203 17:13:42.897528 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:42Z","lastTransitionTime":"2025-12-03T17:13:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.001066 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.001145 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.001157 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.001186 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.001201 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:43Z","lastTransitionTime":"2025-12-03T17:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.103448 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.103495 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.103513 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.103533 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.103547 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:43Z","lastTransitionTime":"2025-12-03T17:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.206520 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.206608 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.206654 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.206696 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.206720 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:43Z","lastTransitionTime":"2025-12-03T17:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.309685 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.309731 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.309743 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.309758 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.309767 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:43Z","lastTransitionTime":"2025-12-03T17:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.413247 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.413311 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.413328 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.413354 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.413371 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:43Z","lastTransitionTime":"2025-12-03T17:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.515891 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.515933 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.515960 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.515976 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.515985 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:43Z","lastTransitionTime":"2025-12-03T17:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.619195 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.619323 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.619345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.619371 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.619389 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:43Z","lastTransitionTime":"2025-12-03T17:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.721997 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.722078 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.722092 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.722111 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.722123 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:43Z","lastTransitionTime":"2025-12-03T17:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.765192 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:43 crc kubenswrapper[4787]: E1203 17:13:43.765398 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.788964 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:43Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.807300 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:43Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.825324 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.825651 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.825777 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.825868 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.825959 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:43Z","lastTransitionTime":"2025-12-03T17:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.828287 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:43Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.854883 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:43Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.870668 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:43Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.890051 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:43Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.904796 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:43Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.920344 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:43Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.928671 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.928733 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.928751 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.928777 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.928795 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:43Z","lastTransitionTime":"2025-12-03T17:13:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.939696 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:43Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.968478 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b886551fd975ca03333f1fea4a4bcfbe92b444de1f51c8c9b30e5cce1602d580\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:33Z\\\",\\\"message\\\":\\\"dler 5 for removal\\\\nI1203 17:13:33.094104 6073 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.094122 6073 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1203 17:13:33.093989 6073 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1203 17:13:33.094057 6073 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094269 6073 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1203 17:13:33.094270 6073 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1203 17:13:33.093998 6073 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1203 17:13:33.094112 6073 factory.go:656] Stopping watch factory\\\\nI1203 17:13:33.094717 6073 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1203 17:13:33.095182 6073 reflector.go:311] Stopping reflector *v1.Service (0s) from k8s.io/client-go/informers/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:29Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\" fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879315 6249 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 17:13:35.879331 6249 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 17:13:35.879422 6249 factory.go:656] Stopping watch factory\\\\nI1203 17:13:35.879438 6249 ovnkube.go:599] Stopped ovnkube\\\\nI1203 17:13:35.879398 6249 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879469 6249 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 17:13:35.879479 6249 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:13:35.879572 6249 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:43Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:43 crc kubenswrapper[4787]: I1203 17:13:43.988533 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:43Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.003979 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:44Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.019528 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:44Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.032409 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.032452 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.032464 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.032483 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.032497 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:44Z","lastTransitionTime":"2025-12-03T17:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.038403 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:44Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.064488 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:44Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.081777 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:44Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.098659 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:44Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.134984 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.135057 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.135080 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.135104 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.135117 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:44Z","lastTransitionTime":"2025-12-03T17:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.177747 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:44 crc kubenswrapper[4787]: E1203 17:13:44.177906 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:44 crc kubenswrapper[4787]: E1203 17:13:44.177970 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs podName:50c2569b-2a14-4112-82e4-afc683aa36a7 nodeName:}" failed. No retries permitted until 2025-12-03 17:13:52.1779539 +0000 UTC m=+48.995424859 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs") pod "network-metrics-daemon-m9tr6" (UID: "50c2569b-2a14-4112-82e4-afc683aa36a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.237253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.237325 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.237348 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.237375 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.237397 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:44Z","lastTransitionTime":"2025-12-03T17:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.339189 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.339240 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.339251 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.339267 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.339279 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:44Z","lastTransitionTime":"2025-12-03T17:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.442074 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.442160 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.442183 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.442225 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.442256 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:44Z","lastTransitionTime":"2025-12-03T17:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.546890 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.546954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.546969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.546990 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.547008 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:44Z","lastTransitionTime":"2025-12-03T17:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.649935 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.649969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.649985 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.650001 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.650011 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:44Z","lastTransitionTime":"2025-12-03T17:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.752852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.753086 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.753147 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.753222 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.753315 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:44Z","lastTransitionTime":"2025-12-03T17:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.765368 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.765493 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:44 crc kubenswrapper[4787]: E1203 17:13:44.765653 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:44 crc kubenswrapper[4787]: E1203 17:13:44.765758 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.765951 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:44 crc kubenswrapper[4787]: E1203 17:13:44.766216 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.856315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.856379 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.856399 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.856424 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.856442 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:44Z","lastTransitionTime":"2025-12-03T17:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.958974 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.959011 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.959041 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.959057 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:44 crc kubenswrapper[4787]: I1203 17:13:44.959068 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:44Z","lastTransitionTime":"2025-12-03T17:13:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.062304 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.062391 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.062417 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.062446 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.062467 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.164413 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.164474 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.164508 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.164539 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.164559 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.268045 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.268094 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.268103 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.268117 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.268126 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.370998 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.371053 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.371063 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.371075 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.371083 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.473614 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.473692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.473717 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.473753 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.473777 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.576561 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.576601 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.576609 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.576624 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.576634 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.679389 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.679474 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.679497 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.679526 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.679547 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.722408 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.722471 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.722495 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.722524 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.722546 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: E1203 17:13:45.743905 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:45Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.748864 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.748922 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.748944 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.748972 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.748996 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.765540 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:45 crc kubenswrapper[4787]: E1203 17:13:45.765774 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:45 crc kubenswrapper[4787]: E1203 17:13:45.771996 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:45Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.776721 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.776761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.776779 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.776802 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.776820 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: E1203 17:13:45.797086 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:45Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.800908 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.800962 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.800981 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.801002 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.801040 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: E1203 17:13:45.818393 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:45Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.822467 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.822544 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.822566 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.822592 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.822613 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: E1203 17:13:45.840609 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:45Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:45Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:45 crc kubenswrapper[4787]: E1203 17:13:45.840831 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.842442 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.842512 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.842536 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.842563 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.842584 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.929814 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.930695 4787 scope.go:117] "RemoveContainer" containerID="0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679" Dec 03 17:13:45 crc kubenswrapper[4787]: E1203 17:13:45.932200 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.944176 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:45Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.944810 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.944844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.944855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.944871 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.944883 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:45Z","lastTransitionTime":"2025-12-03T17:13:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.958201 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:45Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.975687 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:45Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.986215 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:45Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:45 crc kubenswrapper[4787]: I1203 17:13:45.996217 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:45Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.005328 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.018794 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.027972 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.039984 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.046938 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.047116 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.047149 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.047187 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.047213 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:46Z","lastTransitionTime":"2025-12-03T17:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.052874 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.073643 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\" fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879315 6249 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 17:13:35.879331 6249 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 17:13:35.879422 6249 factory.go:656] Stopping watch factory\\\\nI1203 17:13:35.879438 6249 ovnkube.go:599] Stopped ovnkube\\\\nI1203 17:13:35.879398 6249 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879469 6249 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 17:13:35.879479 6249 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:13:35.879572 6249 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.086739 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.105972 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.118347 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.131002 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.145101 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.150154 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.150195 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.150210 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.150230 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.150243 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:46Z","lastTransitionTime":"2025-12-03T17:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.163976 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:46Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.253737 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.253814 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.253827 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.253846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.253860 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:46Z","lastTransitionTime":"2025-12-03T17:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.357125 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.357192 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.357210 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.357233 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.357247 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:46Z","lastTransitionTime":"2025-12-03T17:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.460558 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.460632 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.460655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.460685 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.460709 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:46Z","lastTransitionTime":"2025-12-03T17:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.564375 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.564456 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.564480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.564511 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.564534 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:46Z","lastTransitionTime":"2025-12-03T17:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.667783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.667869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.667892 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.667918 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.667940 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:46Z","lastTransitionTime":"2025-12-03T17:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.765081 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.765115 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:46 crc kubenswrapper[4787]: E1203 17:13:46.765207 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.765081 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:46 crc kubenswrapper[4787]: E1203 17:13:46.765318 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:46 crc kubenswrapper[4787]: E1203 17:13:46.765376 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.770993 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.771076 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.771089 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.771105 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.771116 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:46Z","lastTransitionTime":"2025-12-03T17:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.873829 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.873887 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.873898 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.873917 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.873998 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:46Z","lastTransitionTime":"2025-12-03T17:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.976606 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.976658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.976667 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.976688 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:46 crc kubenswrapper[4787]: I1203 17:13:46.976700 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:46Z","lastTransitionTime":"2025-12-03T17:13:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.079573 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.079619 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.079631 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.079653 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.079666 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:47Z","lastTransitionTime":"2025-12-03T17:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.183126 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.183197 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.183217 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.183245 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.183266 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:47Z","lastTransitionTime":"2025-12-03T17:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.286443 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.286498 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.286515 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.286534 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.286547 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:47Z","lastTransitionTime":"2025-12-03T17:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.389427 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.389490 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.389509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.389536 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.389557 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:47Z","lastTransitionTime":"2025-12-03T17:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.492806 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.492869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.492887 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.492910 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.492925 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:47Z","lastTransitionTime":"2025-12-03T17:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.596328 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.596402 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.596424 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.596454 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.596477 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:47Z","lastTransitionTime":"2025-12-03T17:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.699498 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.699563 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.699580 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.699605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.699622 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:47Z","lastTransitionTime":"2025-12-03T17:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.765775 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:47 crc kubenswrapper[4787]: E1203 17:13:47.766012 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.803592 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.803664 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.803688 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.803717 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.803738 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:47Z","lastTransitionTime":"2025-12-03T17:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.906520 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.906571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.906583 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.906601 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:47 crc kubenswrapper[4787]: I1203 17:13:47.906613 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:47Z","lastTransitionTime":"2025-12-03T17:13:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.009881 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.009938 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.009959 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.009987 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.010008 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:48Z","lastTransitionTime":"2025-12-03T17:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.113323 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.113384 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.113402 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.113426 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.113444 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:48Z","lastTransitionTime":"2025-12-03T17:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.216394 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.216437 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.216453 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.216479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.216496 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:48Z","lastTransitionTime":"2025-12-03T17:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.319314 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.319385 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.319422 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.319452 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.319478 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:48Z","lastTransitionTime":"2025-12-03T17:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.422328 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.422434 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.422455 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.422480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.422498 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:48Z","lastTransitionTime":"2025-12-03T17:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.525725 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.525783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.525806 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.525834 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.525857 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:48Z","lastTransitionTime":"2025-12-03T17:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.629343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.629390 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.629407 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.629429 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.629448 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:48Z","lastTransitionTime":"2025-12-03T17:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.732465 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.732517 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.732529 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.732551 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.732563 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:48Z","lastTransitionTime":"2025-12-03T17:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.765226 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.765303 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.765250 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:48 crc kubenswrapper[4787]: E1203 17:13:48.765478 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:48 crc kubenswrapper[4787]: E1203 17:13:48.765599 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:48 crc kubenswrapper[4787]: E1203 17:13:48.765726 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.834697 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.834756 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.834780 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.834809 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.834831 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:48Z","lastTransitionTime":"2025-12-03T17:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.937068 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.937134 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.937163 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.937192 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:48 crc kubenswrapper[4787]: I1203 17:13:48.937214 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:48Z","lastTransitionTime":"2025-12-03T17:13:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.039415 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.039486 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.039510 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.039544 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.039566 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:49Z","lastTransitionTime":"2025-12-03T17:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.142545 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.142634 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.142668 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.142691 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.142705 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:49Z","lastTransitionTime":"2025-12-03T17:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.246152 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.246204 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.246217 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.246235 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.246247 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:49Z","lastTransitionTime":"2025-12-03T17:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.348673 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.348733 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.348752 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.348778 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.348798 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:49Z","lastTransitionTime":"2025-12-03T17:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.451782 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.451822 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.451832 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.451849 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.451859 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:49Z","lastTransitionTime":"2025-12-03T17:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.554148 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.554207 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.554217 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.554231 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.554242 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:49Z","lastTransitionTime":"2025-12-03T17:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.656286 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.656350 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.656368 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.656392 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.656410 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:49Z","lastTransitionTime":"2025-12-03T17:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.758379 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.758412 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.758421 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.758435 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.758445 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:49Z","lastTransitionTime":"2025-12-03T17:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.765257 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:49 crc kubenswrapper[4787]: E1203 17:13:49.765366 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.860277 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.860340 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.860353 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.860371 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.860382 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:49Z","lastTransitionTime":"2025-12-03T17:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.963803 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.963921 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.963945 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.963969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:49 crc kubenswrapper[4787]: I1203 17:13:49.963986 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:49Z","lastTransitionTime":"2025-12-03T17:13:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.073088 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.073126 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.073135 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.073148 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.073158 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:50Z","lastTransitionTime":"2025-12-03T17:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.175920 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.175975 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.175994 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.176050 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.176069 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:50Z","lastTransitionTime":"2025-12-03T17:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.278675 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.278753 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.278776 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.278810 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.278833 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:50Z","lastTransitionTime":"2025-12-03T17:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.331657 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.344075 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.348666 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.371580 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.382190 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.382308 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.382326 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.382348 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.382365 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:50Z","lastTransitionTime":"2025-12-03T17:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.396226 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.412581 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.431879 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.445785 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.464407 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.481495 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.485081 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.485126 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.485141 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.485162 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.485194 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:50Z","lastTransitionTime":"2025-12-03T17:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.495682 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.517110 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\" fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879315 6249 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 17:13:35.879331 6249 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 17:13:35.879422 6249 factory.go:656] Stopping watch factory\\\\nI1203 17:13:35.879438 6249 ovnkube.go:599] Stopped ovnkube\\\\nI1203 17:13:35.879398 6249 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879469 6249 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 17:13:35.879479 6249 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:13:35.879572 6249 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.527095 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.539579 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.549810 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.561745 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.577854 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.586951 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.586990 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.586999 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.587030 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.587039 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:50Z","lastTransitionTime":"2025-12-03T17:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.593300 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.607316 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:50Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.689049 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.689104 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.689117 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.689136 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.689150 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:50Z","lastTransitionTime":"2025-12-03T17:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.765344 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.765402 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.765476 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:50 crc kubenswrapper[4787]: E1203 17:13:50.765969 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:50 crc kubenswrapper[4787]: E1203 17:13:50.765791 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:50 crc kubenswrapper[4787]: E1203 17:13:50.766153 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.792619 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.792673 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.792686 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.792708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.792720 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:50Z","lastTransitionTime":"2025-12-03T17:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.895723 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.895790 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.895813 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.895844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.895867 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:50Z","lastTransitionTime":"2025-12-03T17:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.999060 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.999132 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.999154 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.999185 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:50 crc kubenswrapper[4787]: I1203 17:13:50.999208 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:50Z","lastTransitionTime":"2025-12-03T17:13:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.102059 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.102131 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.102156 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.102187 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.102208 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:51Z","lastTransitionTime":"2025-12-03T17:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.205269 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.205357 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.205381 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.205412 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.205437 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:51Z","lastTransitionTime":"2025-12-03T17:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.308691 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.308750 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.308761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.308782 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.308791 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:51Z","lastTransitionTime":"2025-12-03T17:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.412009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.412165 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.412191 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.412221 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.412242 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:51Z","lastTransitionTime":"2025-12-03T17:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.519058 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.519127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.519146 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.519174 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.519203 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:51Z","lastTransitionTime":"2025-12-03T17:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.622574 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.622647 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.622670 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.622702 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.622726 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:51Z","lastTransitionTime":"2025-12-03T17:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.726570 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.726626 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.726673 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.726698 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.726715 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:51Z","lastTransitionTime":"2025-12-03T17:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.766067 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:51 crc kubenswrapper[4787]: E1203 17:13:51.766276 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.829902 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.829951 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.829963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.829980 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.829993 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:51Z","lastTransitionTime":"2025-12-03T17:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.933485 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.933576 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.933594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.933619 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:51 crc kubenswrapper[4787]: I1203 17:13:51.933636 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:51Z","lastTransitionTime":"2025-12-03T17:13:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.036296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.036334 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.036344 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.036361 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.036374 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:52Z","lastTransitionTime":"2025-12-03T17:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.139591 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.139644 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.139663 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.139688 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.139705 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:52Z","lastTransitionTime":"2025-12-03T17:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.242062 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.242104 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.242116 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.242134 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.242145 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:52Z","lastTransitionTime":"2025-12-03T17:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.263657 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:52 crc kubenswrapper[4787]: E1203 17:13:52.263798 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:52 crc kubenswrapper[4787]: E1203 17:13:52.263862 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs podName:50c2569b-2a14-4112-82e4-afc683aa36a7 nodeName:}" failed. No retries permitted until 2025-12-03 17:14:08.263844405 +0000 UTC m=+65.081315384 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs") pod "network-metrics-daemon-m9tr6" (UID: "50c2569b-2a14-4112-82e4-afc683aa36a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.344464 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.344517 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.344534 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.344555 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.344570 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:52Z","lastTransitionTime":"2025-12-03T17:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.447936 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.447978 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.447992 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.448011 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.448052 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:52Z","lastTransitionTime":"2025-12-03T17:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.550324 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.550373 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.550389 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.550406 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.550419 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:52Z","lastTransitionTime":"2025-12-03T17:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.652615 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.652683 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.652708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.652737 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.652759 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:52Z","lastTransitionTime":"2025-12-03T17:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.756517 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.756668 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.756707 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.756736 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.756758 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:52Z","lastTransitionTime":"2025-12-03T17:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.765457 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.765502 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.765539 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:52 crc kubenswrapper[4787]: E1203 17:13:52.765599 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:52 crc kubenswrapper[4787]: E1203 17:13:52.765732 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:52 crc kubenswrapper[4787]: E1203 17:13:52.765852 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.859525 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.859599 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.859622 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.859651 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.859672 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:52Z","lastTransitionTime":"2025-12-03T17:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.961994 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.962126 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.962161 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.962207 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:52 crc kubenswrapper[4787]: I1203 17:13:52.962228 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:52Z","lastTransitionTime":"2025-12-03T17:13:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.065976 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.066118 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.066141 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.066173 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.066197 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:53Z","lastTransitionTime":"2025-12-03T17:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.169954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.170054 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.170073 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.170100 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.170128 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:53Z","lastTransitionTime":"2025-12-03T17:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.273739 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.273823 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.273835 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.273852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.273862 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:53Z","lastTransitionTime":"2025-12-03T17:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.376261 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.376312 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.376326 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.376348 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.376361 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:53Z","lastTransitionTime":"2025-12-03T17:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.478842 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.478911 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.478934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.478965 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.478988 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:53Z","lastTransitionTime":"2025-12-03T17:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.581910 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.581950 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.581966 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.581987 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.582004 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:53Z","lastTransitionTime":"2025-12-03T17:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.685935 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.686008 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.686063 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.686123 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.686148 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:53Z","lastTransitionTime":"2025-12-03T17:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.765322 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:53 crc kubenswrapper[4787]: E1203 17:13:53.765488 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.779798 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.789696 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.789775 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.789791 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.789811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.789825 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:53Z","lastTransitionTime":"2025-12-03T17:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.801271 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.815736 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.836473 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.855001 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.872281 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.892627 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.892689 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.892711 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.892736 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.892764 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:53Z","lastTransitionTime":"2025-12-03T17:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.897092 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.914228 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.931514 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.942803 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.957880 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.969143 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.983862 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.995813 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.995851 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.995862 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.995877 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.995889 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:53Z","lastTransitionTime":"2025-12-03T17:13:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:53 crc kubenswrapper[4787]: I1203 17:13:53.997145 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:53Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.021910 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\" fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879315 6249 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 17:13:35.879331 6249 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 17:13:35.879422 6249 factory.go:656] Stopping watch factory\\\\nI1203 17:13:35.879438 6249 ovnkube.go:599] Stopped ovnkube\\\\nI1203 17:13:35.879398 6249 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879469 6249 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 17:13:35.879479 6249 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:13:35.879572 6249 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:54Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.037829 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:54Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.054094 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:54Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.066595 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:54Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.099137 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.099164 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.099174 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.099210 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.099221 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:54Z","lastTransitionTime":"2025-12-03T17:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.202622 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.202700 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.202718 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.202744 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.202762 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:54Z","lastTransitionTime":"2025-12-03T17:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.305764 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.305815 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.305826 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.305843 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.305855 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:54Z","lastTransitionTime":"2025-12-03T17:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.408885 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.408956 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.408978 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.409010 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.409073 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:54Z","lastTransitionTime":"2025-12-03T17:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.512245 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.512287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.512296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.512312 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.512323 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:54Z","lastTransitionTime":"2025-12-03T17:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.584322 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.584478 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.584571 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.584652 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:14:26.584614877 +0000 UTC m=+83.402085876 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.584677 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.584756 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:14:26.58473565 +0000 UTC m=+83.402206649 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.584658 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.584826 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:14:26.584815543 +0000 UTC m=+83.402286502 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.614299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.614361 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.614379 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.614407 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.614424 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:54Z","lastTransitionTime":"2025-12-03T17:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.685842 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.685876 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.686004 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.686040 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.686051 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.686089 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 17:14:26.686077567 +0000 UTC m=+83.503548526 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.686093 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.686141 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.686161 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.686706 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 17:14:26.686230271 +0000 UTC m=+83.503701260 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.716605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.716671 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.716687 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.716707 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.716719 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:54Z","lastTransitionTime":"2025-12-03T17:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.765433 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.765465 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.765603 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.765637 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.765781 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:54 crc kubenswrapper[4787]: E1203 17:13:54.765929 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.820065 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.820146 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.820166 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.820191 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.820211 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:54Z","lastTransitionTime":"2025-12-03T17:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.922893 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.922938 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.922957 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.922982 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:54 crc kubenswrapper[4787]: I1203 17:13:54.923000 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:54Z","lastTransitionTime":"2025-12-03T17:13:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.025827 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.025934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.025955 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.025986 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.026008 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.129415 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.129466 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.129483 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.129507 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.129523 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.232924 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.232958 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.232966 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.232979 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.232988 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.335292 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.335359 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.335382 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.335412 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.335435 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.438228 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.438291 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.438312 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.438335 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.438352 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.541261 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.541325 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.541345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.541369 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.541387 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.644066 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.644128 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.644151 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.644181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.644203 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.746943 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.747000 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.747078 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.747111 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.747133 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.765865 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:55 crc kubenswrapper[4787]: E1203 17:13:55.766046 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.850341 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.850687 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.850700 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.850719 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.850730 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.916430 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.916465 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.916476 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.916499 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.916511 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: E1203 17:13:55.930520 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:55Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.934732 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.934768 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.934777 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.934793 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.934804 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: E1203 17:13:55.948613 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:55Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.953050 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.953175 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.953253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.953350 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.953430 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: E1203 17:13:55.967624 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:55Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.971892 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.972066 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.972152 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.972250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.972344 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:55 crc kubenswrapper[4787]: E1203 17:13:55.988192 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:55Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.992159 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.992225 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.992246 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.992274 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:55 crc kubenswrapper[4787]: I1203 17:13:55.992293 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:55Z","lastTransitionTime":"2025-12-03T17:13:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:56 crc kubenswrapper[4787]: E1203 17:13:56.004251 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:55Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:13:56Z is after 2025-08-24T17:21:41Z" Dec 03 17:13:56 crc kubenswrapper[4787]: E1203 17:13:56.004360 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.005866 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.005903 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.005911 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.005926 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.005935 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:56Z","lastTransitionTime":"2025-12-03T17:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.108341 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.108418 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.108441 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.108475 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.108498 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:56Z","lastTransitionTime":"2025-12-03T17:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.212216 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.212270 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.212288 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.212312 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.212330 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:56Z","lastTransitionTime":"2025-12-03T17:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.316314 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.316402 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.316421 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.316444 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.316461 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:56Z","lastTransitionTime":"2025-12-03T17:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.419409 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.419459 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.419506 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.419550 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.419568 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:56Z","lastTransitionTime":"2025-12-03T17:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.522735 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.523135 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.523343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.523600 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.523885 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:56Z","lastTransitionTime":"2025-12-03T17:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.627454 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.627811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.628315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.628571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.628788 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:56Z","lastTransitionTime":"2025-12-03T17:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.731934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.731992 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.732008 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.732070 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.732092 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:56Z","lastTransitionTime":"2025-12-03T17:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.765638 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.765662 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:56 crc kubenswrapper[4787]: E1203 17:13:56.765752 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.765789 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:56 crc kubenswrapper[4787]: E1203 17:13:56.765840 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:56 crc kubenswrapper[4787]: E1203 17:13:56.765881 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.833966 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.834004 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.834038 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.834064 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.834075 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:56Z","lastTransitionTime":"2025-12-03T17:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.936948 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.936994 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.937005 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.937037 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:56 crc kubenswrapper[4787]: I1203 17:13:56.937050 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:56Z","lastTransitionTime":"2025-12-03T17:13:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.039480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.039528 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.039542 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.039557 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.039570 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:57Z","lastTransitionTime":"2025-12-03T17:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.142433 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.142487 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.142506 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.142535 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.142555 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:57Z","lastTransitionTime":"2025-12-03T17:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.249804 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.249855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.249865 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.249883 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.249894 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:57Z","lastTransitionTime":"2025-12-03T17:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.352450 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.352509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.352527 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.352553 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.352569 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:57Z","lastTransitionTime":"2025-12-03T17:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.454833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.454880 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.454893 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.454910 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.454923 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:57Z","lastTransitionTime":"2025-12-03T17:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.557314 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.557384 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.557401 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.557426 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.557445 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:57Z","lastTransitionTime":"2025-12-03T17:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.659806 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.659873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.659890 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.659913 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.659929 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:57Z","lastTransitionTime":"2025-12-03T17:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.763237 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.763302 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.763314 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.763331 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.763342 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:57Z","lastTransitionTime":"2025-12-03T17:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.765494 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:57 crc kubenswrapper[4787]: E1203 17:13:57.765605 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.865784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.866335 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.866472 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.866666 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.866860 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:57Z","lastTransitionTime":"2025-12-03T17:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.969723 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.969795 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.969818 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.969846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:57 crc kubenswrapper[4787]: I1203 17:13:57.969867 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:57Z","lastTransitionTime":"2025-12-03T17:13:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.072343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.072673 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.073011 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.073220 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.073402 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:58Z","lastTransitionTime":"2025-12-03T17:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.175881 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.176089 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.176183 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.176255 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.176311 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:58Z","lastTransitionTime":"2025-12-03T17:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.281430 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.281536 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.281563 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.281599 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.281631 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:58Z","lastTransitionTime":"2025-12-03T17:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.385555 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.385610 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.385636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.385667 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.385688 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:58Z","lastTransitionTime":"2025-12-03T17:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.489610 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.489686 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.489705 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.489730 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.489748 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:58Z","lastTransitionTime":"2025-12-03T17:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.592700 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.592833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.592853 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.592914 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.592934 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:58Z","lastTransitionTime":"2025-12-03T17:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.695486 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.695568 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.695591 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.695617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.695638 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:58Z","lastTransitionTime":"2025-12-03T17:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.764983 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.764983 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:13:58 crc kubenswrapper[4787]: E1203 17:13:58.765238 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:13:58 crc kubenswrapper[4787]: E1203 17:13:58.765341 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.765011 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:13:58 crc kubenswrapper[4787]: E1203 17:13:58.765465 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.798512 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.798565 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.798584 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.798610 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.798630 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:58Z","lastTransitionTime":"2025-12-03T17:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.902370 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.902432 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.902451 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.902476 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:58 crc kubenswrapper[4787]: I1203 17:13:58.902493 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:58Z","lastTransitionTime":"2025-12-03T17:13:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.004936 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.005073 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.005098 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.005129 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.005154 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:59Z","lastTransitionTime":"2025-12-03T17:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.107888 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.107959 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.107981 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.108009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.108070 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:59Z","lastTransitionTime":"2025-12-03T17:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.211301 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.211468 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.211497 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.211524 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.211550 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:59Z","lastTransitionTime":"2025-12-03T17:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.314758 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.314826 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.314863 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.314896 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.314918 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:59Z","lastTransitionTime":"2025-12-03T17:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.417764 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.417822 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.417846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.417876 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.417900 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:59Z","lastTransitionTime":"2025-12-03T17:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.521486 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.521572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.521595 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.521622 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.521644 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:59Z","lastTransitionTime":"2025-12-03T17:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.624915 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.624963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.624978 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.625000 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.625063 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:59Z","lastTransitionTime":"2025-12-03T17:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.728073 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.728134 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.728152 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.728177 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.728194 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:59Z","lastTransitionTime":"2025-12-03T17:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.765306 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:13:59 crc kubenswrapper[4787]: E1203 17:13:59.765581 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.831063 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.831128 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.831144 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.831168 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.831186 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:59Z","lastTransitionTime":"2025-12-03T17:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.934127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.934192 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.934205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.934234 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:13:59 crc kubenswrapper[4787]: I1203 17:13:59.934257 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:13:59Z","lastTransitionTime":"2025-12-03T17:13:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.037610 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.037695 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.037720 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.037769 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.037794 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:00Z","lastTransitionTime":"2025-12-03T17:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.140435 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.140509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.140531 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.140562 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.140582 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:00Z","lastTransitionTime":"2025-12-03T17:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.243991 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.244045 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.244056 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.244072 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.244082 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:00Z","lastTransitionTime":"2025-12-03T17:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.345683 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.345736 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.345756 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.345779 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.345798 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:00Z","lastTransitionTime":"2025-12-03T17:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.448691 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.448732 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.448741 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.448757 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.448766 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:00Z","lastTransitionTime":"2025-12-03T17:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.551745 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.551794 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.551811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.551828 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.551841 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:00Z","lastTransitionTime":"2025-12-03T17:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.658885 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.658951 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.658969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.658993 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.659009 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:00Z","lastTransitionTime":"2025-12-03T17:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.761478 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.761518 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.761543 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.761558 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.761566 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:00Z","lastTransitionTime":"2025-12-03T17:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.764927 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.764927 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.765065 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:00 crc kubenswrapper[4787]: E1203 17:14:00.765188 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:00 crc kubenswrapper[4787]: E1203 17:14:00.765376 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:00 crc kubenswrapper[4787]: E1203 17:14:00.765760 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.765877 4787 scope.go:117] "RemoveContainer" containerID="0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.865165 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.865233 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.865258 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.865290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.865316 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:00Z","lastTransitionTime":"2025-12-03T17:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.968720 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.968805 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.968819 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.968838 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:00 crc kubenswrapper[4787]: I1203 17:14:00.968851 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:00Z","lastTransitionTime":"2025-12-03T17:14:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.072181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.072228 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.072243 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.072264 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.072279 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:01Z","lastTransitionTime":"2025-12-03T17:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.174621 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.174678 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.174694 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.174715 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.174731 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:01Z","lastTransitionTime":"2025-12-03T17:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.184577 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/1.log" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.186795 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657"} Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.187213 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.211049 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.229276 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.243745 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.257247 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.276810 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.277731 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.277787 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.277805 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.277827 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.277840 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:01Z","lastTransitionTime":"2025-12-03T17:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.290724 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.304140 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.315440 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.334108 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.347099 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.362356 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.375362 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.379572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.379601 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.379609 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.379623 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.379633 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:01Z","lastTransitionTime":"2025-12-03T17:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.391921 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.405227 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.418628 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.429980 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.448389 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\" fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879315 6249 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 17:13:35.879331 6249 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 17:13:35.879422 6249 factory.go:656] Stopping watch factory\\\\nI1203 17:13:35.879438 6249 ovnkube.go:599] Stopped ovnkube\\\\nI1203 17:13:35.879398 6249 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879469 6249 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 17:13:35.879479 6249 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:13:35.879572 6249 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:14:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.459315 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:01Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.482211 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.482250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.482259 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.482273 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.482282 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:01Z","lastTransitionTime":"2025-12-03T17:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.585990 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.586045 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.586063 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.586081 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.586092 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:01Z","lastTransitionTime":"2025-12-03T17:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.689290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.689356 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.689371 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.689393 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.689407 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:01Z","lastTransitionTime":"2025-12-03T17:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.766056 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:01 crc kubenswrapper[4787]: E1203 17:14:01.766380 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.792575 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.792636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.792796 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.792841 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.792863 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:01Z","lastTransitionTime":"2025-12-03T17:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.896416 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.896468 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.896479 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.896494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:01 crc kubenswrapper[4787]: I1203 17:14:01.896505 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:01Z","lastTransitionTime":"2025-12-03T17:14:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.000080 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.000155 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.000174 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.000200 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.000220 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:02Z","lastTransitionTime":"2025-12-03T17:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.103483 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.103895 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.103987 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.104197 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.104311 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:02Z","lastTransitionTime":"2025-12-03T17:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.193336 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/2.log" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.194322 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/1.log" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.198047 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657" exitCode=1 Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.198115 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657"} Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.198180 4787 scope.go:117] "RemoveContainer" containerID="0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.199354 4787 scope.go:117] "RemoveContainer" containerID="91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657" Dec 03 17:14:02 crc kubenswrapper[4787]: E1203 17:14:02.199670 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.208055 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.208108 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.208125 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.208148 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.208169 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:02Z","lastTransitionTime":"2025-12-03T17:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.218231 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.238039 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.269778 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.289163 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.306352 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.311282 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.311322 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.311331 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.311345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.311356 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:02Z","lastTransitionTime":"2025-12-03T17:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.322867 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.342704 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.357683 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.373421 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.393222 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.411691 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.413230 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.413283 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.413299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.413325 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.413342 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:02Z","lastTransitionTime":"2025-12-03T17:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.425951 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.440784 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.453444 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.473114 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0eda8e22e4928f19ecd92d69646ce3bd87c0f1ac6e03f8ed6b1ec509fec6a679\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\" fd69::5 neighbor_responder:none reject:true skip_snat:false]} protocol:{GoSet:[tcp]} selection_fields:{GoSet:[]} vips:{GoMap:map[10.217.5.36:443:]}] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {7f9b8f25-db1a-4d02-a423-9afc5c2fb83c}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879315 6249 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1203 17:13:35.879331 6249 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1203 17:13:35.879422 6249 factory.go:656] Stopping watch factory\\\\nI1203 17:13:35.879438 6249 ovnkube.go:599] Stopped ovnkube\\\\nI1203 17:13:35.879398 6249 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Port_Group Row:map[] Rows:[] Columns:[] Mutations:[{Column:ports Mutator:insert Value:{GoSet:[{GoUUID:960d98b2-dc64-4e93-a4b6-9b19847af71e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {c02bd945-d57b-49ff-9cd3-202ed3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:13:35.879469 6249 handler.go:208] Removed *v1.Node event handler 2\\\\nI1203 17:13:35.879479 6249 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:13:35.879572 6249 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:34Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:01Z\\\",\\\"message\\\":\\\"5-0021ba7e4488}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:14:01.614914 6527 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 17:14:01.614935 6527 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.182\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9099, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 17:14:01.614955 6527 services_controller.go:444] Built service openshift-cluster-version/cluster-version-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.614968 6527 services_controller.go:445] Built service openshift-cluster-version/cluster-version-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.615002 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:14:01.615662 6527 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:00Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.484704 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.498067 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.513261 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:02Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.515933 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.515983 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.516001 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.516037 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.516051 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:02Z","lastTransitionTime":"2025-12-03T17:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.618555 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.618605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.618617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.618635 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.618647 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:02Z","lastTransitionTime":"2025-12-03T17:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.721277 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.721342 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.721352 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.721366 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.721381 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:02Z","lastTransitionTime":"2025-12-03T17:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.765331 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.765533 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:02 crc kubenswrapper[4787]: E1203 17:14:02.765546 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:02 crc kubenswrapper[4787]: E1203 17:14:02.765596 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.765623 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:02 crc kubenswrapper[4787]: E1203 17:14:02.765667 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.823990 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.824084 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.824101 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.824125 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.824147 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:02Z","lastTransitionTime":"2025-12-03T17:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.927182 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.927245 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.927263 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.927287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:02 crc kubenswrapper[4787]: I1203 17:14:02.927303 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:02Z","lastTransitionTime":"2025-12-03T17:14:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.029219 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.029252 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.029260 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.029274 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.029285 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:03Z","lastTransitionTime":"2025-12-03T17:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.132341 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.132379 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.132389 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.132403 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.132413 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:03Z","lastTransitionTime":"2025-12-03T17:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.203266 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/2.log" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.210179 4787 scope.go:117] "RemoveContainer" containerID="91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657" Dec 03 17:14:03 crc kubenswrapper[4787]: E1203 17:14:03.213079 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.225775 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.236210 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.236268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.236280 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.236301 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.236318 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:03Z","lastTransitionTime":"2025-12-03T17:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.244421 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.264939 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.297121 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:01Z\\\",\\\"message\\\":\\\"5-0021ba7e4488}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:14:01.614914 6527 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 17:14:01.614935 6527 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.182\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9099, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 17:14:01.614955 6527 services_controller.go:444] Built service openshift-cluster-version/cluster-version-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.614968 6527 services_controller.go:445] Built service openshift-cluster-version/cluster-version-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.615002 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:14:01.615662 6527 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.313851 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.332437 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.338651 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.338686 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.338705 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.338726 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.338742 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:03Z","lastTransitionTime":"2025-12-03T17:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.349223 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.371916 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.389983 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.418818 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.435592 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.441455 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.441627 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.441849 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.442006 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.442161 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:03Z","lastTransitionTime":"2025-12-03T17:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.450180 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.464179 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.484746 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.497675 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.519165 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.542231 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.544899 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.544961 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.544985 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.545052 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.545077 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:03Z","lastTransitionTime":"2025-12-03T17:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.563087 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.647708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.647765 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.647783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.647812 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.647831 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:03Z","lastTransitionTime":"2025-12-03T17:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.750950 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.750987 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.750999 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.751044 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.751056 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:03Z","lastTransitionTime":"2025-12-03T17:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.765725 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:03 crc kubenswrapper[4787]: E1203 17:14:03.765972 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.791347 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.810013 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.825362 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.851833 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.853490 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.853557 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.853583 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.853612 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.853635 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:03Z","lastTransitionTime":"2025-12-03T17:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.866208 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.880406 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.897462 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.919911 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:01Z\\\",\\\"message\\\":\\\"5-0021ba7e4488}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:14:01.614914 6527 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 17:14:01.614935 6527 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.182\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9099, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 17:14:01.614955 6527 services_controller.go:444] Built service openshift-cluster-version/cluster-version-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.614968 6527 services_controller.go:445] Built service openshift-cluster-version/cluster-version-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.615002 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:14:01.615662 6527 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.933487 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.944786 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.956413 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.956461 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.956490 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.956517 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.956533 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:03Z","lastTransitionTime":"2025-12-03T17:14:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.958711 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.975156 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:03 crc kubenswrapper[4787]: I1203 17:14:03.990614 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:03Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.006880 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:04Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.022776 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:04Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.037117 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:04Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.051255 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:04Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.059120 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.059309 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.059396 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.059496 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.059585 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:04Z","lastTransitionTime":"2025-12-03T17:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.070326 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:04Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.162267 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.162574 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.162663 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.162756 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.162850 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:04Z","lastTransitionTime":"2025-12-03T17:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.265226 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.265285 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.265302 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.265329 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.265345 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:04Z","lastTransitionTime":"2025-12-03T17:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.370152 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.370221 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.370241 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.370267 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.370295 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:04Z","lastTransitionTime":"2025-12-03T17:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.473429 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.473477 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.473491 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.473508 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.473521 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:04Z","lastTransitionTime":"2025-12-03T17:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.575746 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.575810 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.575828 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.575855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.575871 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:04Z","lastTransitionTime":"2025-12-03T17:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.678730 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.678800 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.678819 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.678844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.678863 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:04Z","lastTransitionTime":"2025-12-03T17:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.765976 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.766065 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:04 crc kubenswrapper[4787]: E1203 17:14:04.766135 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.766150 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:04 crc kubenswrapper[4787]: E1203 17:14:04.766264 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:04 crc kubenswrapper[4787]: E1203 17:14:04.766389 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.781898 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.781966 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.781978 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.781998 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.782010 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:04Z","lastTransitionTime":"2025-12-03T17:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.884998 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.885048 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.885057 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.885071 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.885082 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:04Z","lastTransitionTime":"2025-12-03T17:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.987562 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.987606 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.987624 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.987641 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:04 crc kubenswrapper[4787]: I1203 17:14:04.987656 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:04Z","lastTransitionTime":"2025-12-03T17:14:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.090608 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.090684 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.090709 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.090737 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.090755 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:05Z","lastTransitionTime":"2025-12-03T17:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.193910 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.193965 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.193982 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.194005 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.194063 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:05Z","lastTransitionTime":"2025-12-03T17:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.297169 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.297584 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.297723 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.297901 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.298116 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:05Z","lastTransitionTime":"2025-12-03T17:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.400873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.401171 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.401254 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.401346 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.401439 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:05Z","lastTransitionTime":"2025-12-03T17:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.503848 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.504203 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.504213 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.504226 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.504238 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:05Z","lastTransitionTime":"2025-12-03T17:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.606362 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.606422 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.606438 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.606466 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.606483 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:05Z","lastTransitionTime":"2025-12-03T17:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.708655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.709011 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.709226 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.709523 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.709720 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:05Z","lastTransitionTime":"2025-12-03T17:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.765481 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:05 crc kubenswrapper[4787]: E1203 17:14:05.765683 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.812551 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.812616 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.812629 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.812646 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.812660 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:05Z","lastTransitionTime":"2025-12-03T17:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.915992 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.916057 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.916070 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.916089 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:05 crc kubenswrapper[4787]: I1203 17:14:05.916101 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:05Z","lastTransitionTime":"2025-12-03T17:14:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.018733 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.018781 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.018796 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.018814 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.018826 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.120641 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.120690 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.120698 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.120712 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.120722 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.223249 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.223290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.223303 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.223320 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.223332 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.325183 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.325536 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.325694 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.325836 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.325965 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: E1203 17:14:06.346453 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:06Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.351683 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.351903 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.352101 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.352291 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.352450 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: E1203 17:14:06.374519 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:06Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.381673 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.381710 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.381718 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.381734 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.381747 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: E1203 17:14:06.395862 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:06Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.399294 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.399325 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.399335 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.399347 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.399357 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: E1203 17:14:06.417684 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:06Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.423298 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.423342 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.423352 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.423371 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.423383 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: E1203 17:14:06.437665 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:06Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:06Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:06 crc kubenswrapper[4787]: E1203 17:14:06.437911 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.439623 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.439670 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.439686 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.439708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.439724 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.542844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.542885 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.542894 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.542911 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.542920 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.645967 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.646092 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.646166 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.646279 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.646307 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.748762 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.748820 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.748835 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.748860 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.748878 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.765295 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.765347 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:06 crc kubenswrapper[4787]: E1203 17:14:06.765452 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.765295 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:06 crc kubenswrapper[4787]: E1203 17:14:06.765643 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:06 crc kubenswrapper[4787]: E1203 17:14:06.765681 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.852241 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.852300 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.852317 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.852340 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.852356 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.954938 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.954968 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.954979 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.954995 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:06 crc kubenswrapper[4787]: I1203 17:14:06.955059 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:06Z","lastTransitionTime":"2025-12-03T17:14:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.056895 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.056941 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.056956 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.056976 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.056989 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:07Z","lastTransitionTime":"2025-12-03T17:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.159406 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.159460 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.159473 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.159493 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.159505 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:07Z","lastTransitionTime":"2025-12-03T17:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.262093 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.262127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.262139 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.262156 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.262170 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:07Z","lastTransitionTime":"2025-12-03T17:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.364814 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.364866 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.364902 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.364923 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.364936 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:07Z","lastTransitionTime":"2025-12-03T17:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.467693 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.467722 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.467730 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.467745 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.467754 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:07Z","lastTransitionTime":"2025-12-03T17:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.570291 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.570354 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.570367 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.570384 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.570395 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:07Z","lastTransitionTime":"2025-12-03T17:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.693680 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.693733 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.693751 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.693774 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.693790 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:07Z","lastTransitionTime":"2025-12-03T17:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.765466 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:07 crc kubenswrapper[4787]: E1203 17:14:07.765615 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.796506 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.796591 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.796625 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.796655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.796680 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:07Z","lastTransitionTime":"2025-12-03T17:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.899258 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.899315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.899327 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.899347 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:07 crc kubenswrapper[4787]: I1203 17:14:07.899358 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:07Z","lastTransitionTime":"2025-12-03T17:14:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.003252 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.003295 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.003322 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.003337 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.003349 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:08Z","lastTransitionTime":"2025-12-03T17:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.106139 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.106177 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.106192 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.106214 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.106228 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:08Z","lastTransitionTime":"2025-12-03T17:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.209193 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.209226 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.209234 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.209246 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.209255 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:08Z","lastTransitionTime":"2025-12-03T17:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.311511 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.311578 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.311593 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.311610 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.311624 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:08Z","lastTransitionTime":"2025-12-03T17:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.330499 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:08 crc kubenswrapper[4787]: E1203 17:14:08.330674 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:14:08 crc kubenswrapper[4787]: E1203 17:14:08.330751 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs podName:50c2569b-2a14-4112-82e4-afc683aa36a7 nodeName:}" failed. No retries permitted until 2025-12-03 17:14:40.330731061 +0000 UTC m=+97.148202060 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs") pod "network-metrics-daemon-m9tr6" (UID: "50c2569b-2a14-4112-82e4-afc683aa36a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.415104 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.415152 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.415168 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.415197 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.415210 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:08Z","lastTransitionTime":"2025-12-03T17:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.518231 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.518288 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.518301 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.518319 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.518332 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:08Z","lastTransitionTime":"2025-12-03T17:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.620625 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.620678 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.620690 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.620705 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.620714 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:08Z","lastTransitionTime":"2025-12-03T17:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.723202 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.723237 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.723249 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.723262 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.723271 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:08Z","lastTransitionTime":"2025-12-03T17:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.765672 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.765685 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.765809 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:08 crc kubenswrapper[4787]: E1203 17:14:08.766087 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:08 crc kubenswrapper[4787]: E1203 17:14:08.766292 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:08 crc kubenswrapper[4787]: E1203 17:14:08.766372 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.825086 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.825141 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.825157 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.825179 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.825194 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:08Z","lastTransitionTime":"2025-12-03T17:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.927754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.927804 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.927819 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.927839 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:08 crc kubenswrapper[4787]: I1203 17:14:08.927852 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:08Z","lastTransitionTime":"2025-12-03T17:14:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.030416 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.030444 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.030452 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.030467 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.030477 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:09Z","lastTransitionTime":"2025-12-03T17:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.132784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.132848 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.132862 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.132884 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.132898 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:09Z","lastTransitionTime":"2025-12-03T17:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.234720 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.234766 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.234779 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.234814 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.234824 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:09Z","lastTransitionTime":"2025-12-03T17:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.337335 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.337382 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.337397 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.337414 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.337426 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:09Z","lastTransitionTime":"2025-12-03T17:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.439615 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.439659 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.439670 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.439687 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.439698 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:09Z","lastTransitionTime":"2025-12-03T17:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.542187 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.542243 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.542260 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.542317 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.542345 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:09Z","lastTransitionTime":"2025-12-03T17:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.644811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.644852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.644865 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.644883 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.644895 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:09Z","lastTransitionTime":"2025-12-03T17:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.746907 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.746953 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.746963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.746980 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.746994 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:09Z","lastTransitionTime":"2025-12-03T17:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.765810 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:09 crc kubenswrapper[4787]: E1203 17:14:09.765989 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.849915 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.849953 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.849962 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.849977 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.849988 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:09Z","lastTransitionTime":"2025-12-03T17:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.952519 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.952561 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.952573 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.952590 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:09 crc kubenswrapper[4787]: I1203 17:14:09.952600 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:09Z","lastTransitionTime":"2025-12-03T17:14:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.054719 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.054753 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.054764 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.054779 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.054789 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:10Z","lastTransitionTime":"2025-12-03T17:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.157317 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.157344 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.157365 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.157377 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.157389 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:10Z","lastTransitionTime":"2025-12-03T17:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.230239 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xcz4f_bd5617c0-aa4f-4f21-a131-15af831725ec/kube-multus/0.log" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.230284 4787 generic.go:334] "Generic (PLEG): container finished" podID="bd5617c0-aa4f-4f21-a131-15af831725ec" containerID="d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d" exitCode=1 Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.230309 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xcz4f" event={"ID":"bd5617c0-aa4f-4f21-a131-15af831725ec","Type":"ContainerDied","Data":"d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d"} Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.230622 4787 scope.go:117] "RemoveContainer" containerID="d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.257545 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:10Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:09Z\\\",\\\"message\\\":\\\"2025-12-03T17:13:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9\\\\n2025-12-03T17:13:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9 to /host/opt/cni/bin/\\\\n2025-12-03T17:13:24Z [verbose] multus-daemon started\\\\n2025-12-03T17:13:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T17:14:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.259771 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.259809 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.259822 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.259837 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.259847 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:10Z","lastTransitionTime":"2025-12-03T17:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.274328 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.306967 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.324305 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.335864 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.345501 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.359484 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.364505 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.364540 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.364548 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.364562 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.364572 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:10Z","lastTransitionTime":"2025-12-03T17:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.372002 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.384293 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.397296 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.410900 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.420358 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.431576 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.441556 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.458595 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:01Z\\\",\\\"message\\\":\\\"5-0021ba7e4488}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:14:01.614914 6527 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 17:14:01.614935 6527 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.182\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9099, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 17:14:01.614955 6527 services_controller.go:444] Built service openshift-cluster-version/cluster-version-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.614968 6527 services_controller.go:445] Built service openshift-cluster-version/cluster-version-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.615002 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:14:01.615662 6527 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.467055 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.467102 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.467115 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.467133 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.467146 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:10Z","lastTransitionTime":"2025-12-03T17:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.469795 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.482108 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.493817 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:10Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.570104 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.570151 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.570164 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.570180 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.570193 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:10Z","lastTransitionTime":"2025-12-03T17:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.673108 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.673157 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.673168 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.673185 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.673194 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:10Z","lastTransitionTime":"2025-12-03T17:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.765292 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.765481 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:10 crc kubenswrapper[4787]: E1203 17:14:10.765595 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.765696 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:10 crc kubenswrapper[4787]: E1203 17:14:10.765836 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:10 crc kubenswrapper[4787]: E1203 17:14:10.766061 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.775101 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.775141 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.775153 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.775169 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.775180 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:10Z","lastTransitionTime":"2025-12-03T17:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.877310 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.877350 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.877361 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.877393 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.877405 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:10Z","lastTransitionTime":"2025-12-03T17:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.980204 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.980238 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.980247 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.980262 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:10 crc kubenswrapper[4787]: I1203 17:14:10.980272 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:10Z","lastTransitionTime":"2025-12-03T17:14:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.082617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.082647 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.082655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.082691 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.082700 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:11Z","lastTransitionTime":"2025-12-03T17:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.185469 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.185509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.185518 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.185534 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.185545 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:11Z","lastTransitionTime":"2025-12-03T17:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.234512 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xcz4f_bd5617c0-aa4f-4f21-a131-15af831725ec/kube-multus/0.log" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.234564 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xcz4f" event={"ID":"bd5617c0-aa4f-4f21-a131-15af831725ec","Type":"ContainerStarted","Data":"59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8"} Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.246246 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.261713 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:01Z\\\",\\\"message\\\":\\\"5-0021ba7e4488}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:14:01.614914 6527 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 17:14:01.614935 6527 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.182\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9099, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 17:14:01.614955 6527 services_controller.go:444] Built service openshift-cluster-version/cluster-version-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.614968 6527 services_controller.go:445] Built service openshift-cluster-version/cluster-version-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.615002 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:14:01.615662 6527 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.270316 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.280839 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.287636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.287665 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.287674 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.287687 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.287695 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:11Z","lastTransitionTime":"2025-12-03T17:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.290784 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.302842 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.312311 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.328444 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.340327 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.351230 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:09Z\\\",\\\"message\\\":\\\"2025-12-03T17:13:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9\\\\n2025-12-03T17:13:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9 to /host/opt/cni/bin/\\\\n2025-12-03T17:13:24Z [verbose] multus-daemon started\\\\n2025-12-03T17:13:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T17:14:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:14:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.361782 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.373615 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.386923 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.393449 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.393483 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.393494 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.393508 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.393522 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:11Z","lastTransitionTime":"2025-12-03T17:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.401241 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.411529 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.426473 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.438001 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.453647 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:11Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.496009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.496081 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.496097 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.496120 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.496174 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:11Z","lastTransitionTime":"2025-12-03T17:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.598636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.598685 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.598697 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.598727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.598741 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:11Z","lastTransitionTime":"2025-12-03T17:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.700586 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.700630 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.700641 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.700658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.700675 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:11Z","lastTransitionTime":"2025-12-03T17:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.765903 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:11 crc kubenswrapper[4787]: E1203 17:14:11.766064 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.803119 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.803188 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.803211 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.803238 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.803259 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:11Z","lastTransitionTime":"2025-12-03T17:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.905361 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.905408 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.905418 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.905433 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:11 crc kubenswrapper[4787]: I1203 17:14:11.905444 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:11Z","lastTransitionTime":"2025-12-03T17:14:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.008065 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.008114 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.008126 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.008142 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.008154 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:12Z","lastTransitionTime":"2025-12-03T17:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.110855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.110895 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.110908 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.110924 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.110934 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:12Z","lastTransitionTime":"2025-12-03T17:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.213308 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.213383 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.213406 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.213431 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.213449 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:12Z","lastTransitionTime":"2025-12-03T17:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.315535 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.315584 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.315598 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.315615 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.315628 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:12Z","lastTransitionTime":"2025-12-03T17:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.417768 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.417829 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.417846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.417872 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.417892 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:12Z","lastTransitionTime":"2025-12-03T17:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.519889 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.519923 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.519931 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.519945 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.519954 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:12Z","lastTransitionTime":"2025-12-03T17:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.622826 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.622868 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.622877 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.622891 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.622903 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:12Z","lastTransitionTime":"2025-12-03T17:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.724926 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.724953 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.724961 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.724972 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.724981 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:12Z","lastTransitionTime":"2025-12-03T17:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.765134 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.765154 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.765134 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:12 crc kubenswrapper[4787]: E1203 17:14:12.765265 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:12 crc kubenswrapper[4787]: E1203 17:14:12.765315 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:12 crc kubenswrapper[4787]: E1203 17:14:12.765381 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.827338 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.827409 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.827427 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.827448 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.827460 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:12Z","lastTransitionTime":"2025-12-03T17:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.929809 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.929844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.929852 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.929865 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:12 crc kubenswrapper[4787]: I1203 17:14:12.929874 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:12Z","lastTransitionTime":"2025-12-03T17:14:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.032315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.032368 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.032380 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.032398 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.032409 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:13Z","lastTransitionTime":"2025-12-03T17:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.135064 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.135104 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.135113 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.135128 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.135138 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:13Z","lastTransitionTime":"2025-12-03T17:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.236769 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.236816 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.236828 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.236845 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.236857 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:13Z","lastTransitionTime":"2025-12-03T17:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.339761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.339799 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.339809 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.339826 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.339836 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:13Z","lastTransitionTime":"2025-12-03T17:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.442272 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.442310 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.442321 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.442336 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.442349 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:13Z","lastTransitionTime":"2025-12-03T17:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.545155 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.545216 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.545253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.545290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.545316 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:13Z","lastTransitionTime":"2025-12-03T17:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.647539 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.647599 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.647614 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.647632 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.647646 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:13Z","lastTransitionTime":"2025-12-03T17:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.750330 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.750379 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.750391 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.750409 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.750420 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:13Z","lastTransitionTime":"2025-12-03T17:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.765597 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:13 crc kubenswrapper[4787]: E1203 17:14:13.766090 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.766347 4787 scope.go:117] "RemoveContainer" containerID="91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657" Dec 03 17:14:13 crc kubenswrapper[4787]: E1203 17:14:13.766557 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.784804 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.799784 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.814328 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.827795 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.847341 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:09Z\\\",\\\"message\\\":\\\"2025-12-03T17:13:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9\\\\n2025-12-03T17:13:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9 to /host/opt/cni/bin/\\\\n2025-12-03T17:13:24Z [verbose] multus-daemon started\\\\n2025-12-03T17:13:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T17:14:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:14:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.851764 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.851861 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.851961 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.852070 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.852138 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:13Z","lastTransitionTime":"2025-12-03T17:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.862477 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.886915 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.898624 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.912913 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.923096 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.940503 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.951359 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.954758 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.954813 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.954832 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.954856 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.954913 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:13Z","lastTransitionTime":"2025-12-03T17:14:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.970463 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:13 crc kubenswrapper[4787]: I1203 17:14:13.987587 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:13Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.014691 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:01Z\\\",\\\"message\\\":\\\"5-0021ba7e4488}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:14:01.614914 6527 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 17:14:01.614935 6527 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.182\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9099, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 17:14:01.614955 6527 services_controller.go:444] Built service openshift-cluster-version/cluster-version-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.614968 6527 services_controller.go:445] Built service openshift-cluster-version/cluster-version-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.615002 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:14:01.615662 6527 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:14Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.028513 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:14Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.042036 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:14Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.054152 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:14Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.057056 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.057085 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.057094 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.057111 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.057121 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:14Z","lastTransitionTime":"2025-12-03T17:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.159591 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.159615 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.159623 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.159636 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.159644 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:14Z","lastTransitionTime":"2025-12-03T17:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.261223 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.261272 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.261281 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.261293 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.261302 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:14Z","lastTransitionTime":"2025-12-03T17:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.364733 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.364771 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.364783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.364800 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.364812 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:14Z","lastTransitionTime":"2025-12-03T17:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.467093 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.467130 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.467149 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.467197 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.467231 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:14Z","lastTransitionTime":"2025-12-03T17:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.569411 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.569451 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.569464 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.569480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.569490 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:14Z","lastTransitionTime":"2025-12-03T17:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.671702 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.671730 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.671738 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.671750 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.671758 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:14Z","lastTransitionTime":"2025-12-03T17:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.765174 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.765210 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.765367 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:14 crc kubenswrapper[4787]: E1203 17:14:14.765427 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:14 crc kubenswrapper[4787]: E1203 17:14:14.765566 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:14 crc kubenswrapper[4787]: E1203 17:14:14.765689 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.774113 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.774143 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.774153 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.774166 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.774178 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:14Z","lastTransitionTime":"2025-12-03T17:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.876679 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.876727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.876738 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.876752 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.876763 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:14Z","lastTransitionTime":"2025-12-03T17:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.979849 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.979916 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.979932 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.979967 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:14 crc kubenswrapper[4787]: I1203 17:14:14.979989 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:14Z","lastTransitionTime":"2025-12-03T17:14:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.082285 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.082329 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.082339 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.082354 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.082362 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:15Z","lastTransitionTime":"2025-12-03T17:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.184832 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.184865 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.184873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.184886 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.184895 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:15Z","lastTransitionTime":"2025-12-03T17:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.286692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.286724 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.286731 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.286744 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.286753 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:15Z","lastTransitionTime":"2025-12-03T17:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.389067 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.389102 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.389112 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.389130 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.389143 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:15Z","lastTransitionTime":"2025-12-03T17:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.490748 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.490782 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.490790 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.490804 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.490816 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:15Z","lastTransitionTime":"2025-12-03T17:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.592941 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.592985 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.592999 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.593032 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.593043 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:15Z","lastTransitionTime":"2025-12-03T17:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.695278 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.695334 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.695348 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.695369 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.695381 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:15Z","lastTransitionTime":"2025-12-03T17:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.765291 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:15 crc kubenswrapper[4787]: E1203 17:14:15.765429 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.798399 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.798445 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.798456 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.798470 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.798480 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:15Z","lastTransitionTime":"2025-12-03T17:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.901094 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.901158 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.901181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.901209 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:15 crc kubenswrapper[4787]: I1203 17:14:15.901232 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:15Z","lastTransitionTime":"2025-12-03T17:14:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.003572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.003859 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.003967 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.004083 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.004208 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.106324 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.106745 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.106945 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.107118 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.107242 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.210713 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.210755 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.210766 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.210783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.210795 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.312762 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.312833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.312846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.312866 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.312885 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.445186 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.445225 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.445235 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.445250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.445266 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.547904 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.547963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.547984 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.548005 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.548042 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.650702 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.650748 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.650763 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.650785 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.650800 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.753283 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.753344 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.753367 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.753396 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.753417 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.765516 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.765564 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:16 crc kubenswrapper[4787]: E1203 17:14:16.765647 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.765586 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:16 crc kubenswrapper[4787]: E1203 17:14:16.765743 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:16 crc kubenswrapper[4787]: E1203 17:14:16.765815 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.836795 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.836831 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.836840 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.836854 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.836863 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: E1203 17:14:16.852546 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:16Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.856590 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.856658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.856669 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.856685 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.856695 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: E1203 17:14:16.872199 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:16Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.875587 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.875629 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.875644 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.875667 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.875683 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: E1203 17:14:16.889383 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:16Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.892951 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.893042 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.893069 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.893095 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.893113 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: E1203 17:14:16.906853 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:16Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.910518 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.910569 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.910585 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.910610 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.910624 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:16 crc kubenswrapper[4787]: E1203 17:14:16.925659 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:16Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:16Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:16 crc kubenswrapper[4787]: E1203 17:14:16.925792 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.927200 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.927255 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.927266 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.927279 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:16 crc kubenswrapper[4787]: I1203 17:14:16.927288 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:16Z","lastTransitionTime":"2025-12-03T17:14:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.029315 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.029605 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.029771 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.029926 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.030101 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:17Z","lastTransitionTime":"2025-12-03T17:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.133011 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.133428 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.133684 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.133946 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.134262 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:17Z","lastTransitionTime":"2025-12-03T17:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.237517 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.237572 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.237594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.237689 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.237716 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:17Z","lastTransitionTime":"2025-12-03T17:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.340815 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.340873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.340891 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.340915 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.340932 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:17Z","lastTransitionTime":"2025-12-03T17:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.443857 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.443930 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.443954 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.443978 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.443995 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:17Z","lastTransitionTime":"2025-12-03T17:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.546679 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.546742 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.546754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.546769 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.546781 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:17Z","lastTransitionTime":"2025-12-03T17:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.649433 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.649474 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.649482 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.649498 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.649508 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:17Z","lastTransitionTime":"2025-12-03T17:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.752793 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.752858 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.752876 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.752900 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.752920 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:17Z","lastTransitionTime":"2025-12-03T17:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.765788 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:17 crc kubenswrapper[4787]: E1203 17:14:17.766100 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.778536 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.856231 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.856309 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.856335 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.856364 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.856385 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:17Z","lastTransitionTime":"2025-12-03T17:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.959719 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.959760 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.959773 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.959791 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:17 crc kubenswrapper[4787]: I1203 17:14:17.959803 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:17Z","lastTransitionTime":"2025-12-03T17:14:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.062741 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.062792 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.062810 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.062833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.062850 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:18Z","lastTransitionTime":"2025-12-03T17:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.165615 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.165938 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.166054 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.166146 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.166241 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:18Z","lastTransitionTime":"2025-12-03T17:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.268701 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.268759 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.268781 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.268802 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.268817 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:18Z","lastTransitionTime":"2025-12-03T17:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.371934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.372059 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.372086 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.372116 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.372137 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:18Z","lastTransitionTime":"2025-12-03T17:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.475701 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.475772 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.475791 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.475815 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.475856 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:18Z","lastTransitionTime":"2025-12-03T17:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.578358 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.578394 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.578404 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.578416 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.578424 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:18Z","lastTransitionTime":"2025-12-03T17:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.681634 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.681677 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.681687 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.681701 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.681711 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:18Z","lastTransitionTime":"2025-12-03T17:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.765969 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.766073 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.766097 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:18 crc kubenswrapper[4787]: E1203 17:14:18.766169 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:18 crc kubenswrapper[4787]: E1203 17:14:18.766324 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:18 crc kubenswrapper[4787]: E1203 17:14:18.766631 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.790044 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.790094 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.790107 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.790127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.790143 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:18Z","lastTransitionTime":"2025-12-03T17:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.893295 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.893357 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.893382 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.893412 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.893432 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:18Z","lastTransitionTime":"2025-12-03T17:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.995720 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.996074 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.996176 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.996261 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:18 crc kubenswrapper[4787]: I1203 17:14:18.996336 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:18Z","lastTransitionTime":"2025-12-03T17:14:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.099677 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.099759 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.099796 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.099826 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.099846 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:19Z","lastTransitionTime":"2025-12-03T17:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.203289 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.203343 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.203355 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.203373 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.203387 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:19Z","lastTransitionTime":"2025-12-03T17:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.305899 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.305963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.305988 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.306049 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.306072 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:19Z","lastTransitionTime":"2025-12-03T17:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.408467 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.408566 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.408583 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.408606 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.408622 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:19Z","lastTransitionTime":"2025-12-03T17:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.511095 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.511156 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.511172 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.511196 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.511212 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:19Z","lastTransitionTime":"2025-12-03T17:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.615258 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.615321 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.615344 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.615376 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.615401 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:19Z","lastTransitionTime":"2025-12-03T17:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.718612 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.718691 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.718709 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.718733 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.718753 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:19Z","lastTransitionTime":"2025-12-03T17:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.765272 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:19 crc kubenswrapper[4787]: E1203 17:14:19.765408 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.821950 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.821996 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.822009 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.822041 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.822051 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:19Z","lastTransitionTime":"2025-12-03T17:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.925509 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.925585 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.925599 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.925617 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:19 crc kubenswrapper[4787]: I1203 17:14:19.925653 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:19Z","lastTransitionTime":"2025-12-03T17:14:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.028992 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.029060 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.029071 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.029087 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.029099 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:20Z","lastTransitionTime":"2025-12-03T17:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.133219 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.133260 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.133272 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.133289 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.133300 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:20Z","lastTransitionTime":"2025-12-03T17:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.236096 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.236162 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.236180 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.236205 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.236223 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:20Z","lastTransitionTime":"2025-12-03T17:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.339117 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.339168 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.339185 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.339211 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.339226 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:20Z","lastTransitionTime":"2025-12-03T17:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.441406 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.441465 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.441476 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.441496 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.441511 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:20Z","lastTransitionTime":"2025-12-03T17:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.544459 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.544527 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.544549 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.544576 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.544595 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:20Z","lastTransitionTime":"2025-12-03T17:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.647143 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.647180 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.647189 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.647202 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.647210 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:20Z","lastTransitionTime":"2025-12-03T17:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.749576 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.749613 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.749627 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.749642 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.749651 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:20Z","lastTransitionTime":"2025-12-03T17:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.765569 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.765648 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:20 crc kubenswrapper[4787]: E1203 17:14:20.765701 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:20 crc kubenswrapper[4787]: E1203 17:14:20.765756 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.765882 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:20 crc kubenswrapper[4787]: E1203 17:14:20.766046 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.852383 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.852441 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.852461 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.852487 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.852504 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:20Z","lastTransitionTime":"2025-12-03T17:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.954959 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.955012 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.955038 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.955052 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:20 crc kubenswrapper[4787]: I1203 17:14:20.955062 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:20Z","lastTransitionTime":"2025-12-03T17:14:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.058115 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.058188 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.058211 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.058232 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.058247 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:21Z","lastTransitionTime":"2025-12-03T17:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.161258 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.161342 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.161365 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.161398 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.161419 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:21Z","lastTransitionTime":"2025-12-03T17:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.264613 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.264684 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.264701 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.264723 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.264744 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:21Z","lastTransitionTime":"2025-12-03T17:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.367476 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.367518 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.367530 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.367544 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.367555 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:21Z","lastTransitionTime":"2025-12-03T17:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.470791 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.470856 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.470879 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.470920 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.470957 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:21Z","lastTransitionTime":"2025-12-03T17:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.573907 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.573936 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.573945 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.573959 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.573969 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:21Z","lastTransitionTime":"2025-12-03T17:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.677382 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.677443 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.677463 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.677488 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.677509 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:21Z","lastTransitionTime":"2025-12-03T17:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.765956 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:21 crc kubenswrapper[4787]: E1203 17:14:21.766187 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.781063 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.781134 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.781152 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.781177 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.781195 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:21Z","lastTransitionTime":"2025-12-03T17:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.883743 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.883800 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.883813 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.883834 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.883846 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:21Z","lastTransitionTime":"2025-12-03T17:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.987132 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.987181 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.987193 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.987212 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:21 crc kubenswrapper[4787]: I1203 17:14:21.987223 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:21Z","lastTransitionTime":"2025-12-03T17:14:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.090105 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.090143 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.090154 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.090170 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.090181 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:22Z","lastTransitionTime":"2025-12-03T17:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.192459 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.192496 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.192505 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.192520 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.192529 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:22Z","lastTransitionTime":"2025-12-03T17:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.294959 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.295013 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.295057 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.295077 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.295093 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:22Z","lastTransitionTime":"2025-12-03T17:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.398008 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.398106 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.398123 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.398149 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.398165 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:22Z","lastTransitionTime":"2025-12-03T17:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.501270 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.501314 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.501327 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.501344 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.501356 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:22Z","lastTransitionTime":"2025-12-03T17:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.604752 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.604803 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.604821 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.604845 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.604866 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:22Z","lastTransitionTime":"2025-12-03T17:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.707386 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.707445 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.707470 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.707499 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.707520 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:22Z","lastTransitionTime":"2025-12-03T17:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.764974 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.765085 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.764974 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:22 crc kubenswrapper[4787]: E1203 17:14:22.765192 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:22 crc kubenswrapper[4787]: E1203 17:14:22.765541 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:22 crc kubenswrapper[4787]: E1203 17:14:22.766136 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.811912 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.811973 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.811996 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.812061 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.812087 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:22Z","lastTransitionTime":"2025-12-03T17:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.915797 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.915856 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.915875 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.915900 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:22 crc kubenswrapper[4787]: I1203 17:14:22.915918 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:22Z","lastTransitionTime":"2025-12-03T17:14:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.018300 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.018364 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.018387 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.018415 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.018438 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:23Z","lastTransitionTime":"2025-12-03T17:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.123761 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.123833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.123846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.123869 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.123885 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:23Z","lastTransitionTime":"2025-12-03T17:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.226898 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.226964 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.226981 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.227006 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.227076 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:23Z","lastTransitionTime":"2025-12-03T17:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.330306 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.330405 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.330431 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.330464 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.330488 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:23Z","lastTransitionTime":"2025-12-03T17:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.432886 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.432938 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.432957 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.432982 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.433000 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:23Z","lastTransitionTime":"2025-12-03T17:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.536069 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.536110 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.536122 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.536138 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.536151 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:23Z","lastTransitionTime":"2025-12-03T17:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.640220 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.640296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.640333 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.640362 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.640380 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:23Z","lastTransitionTime":"2025-12-03T17:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.742992 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.743040 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.743050 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.743067 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.743076 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:23Z","lastTransitionTime":"2025-12-03T17:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.765209 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:23 crc kubenswrapper[4787]: E1203 17:14:23.765452 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.779556 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.795401 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.820800 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:01Z\\\",\\\"message\\\":\\\"5-0021ba7e4488}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:14:01.614914 6527 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 17:14:01.614935 6527 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.182\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9099, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 17:14:01.614955 6527 services_controller.go:444] Built service openshift-cluster-version/cluster-version-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.614968 6527 services_controller.go:445] Built service openshift-cluster-version/cluster-version-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.615002 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:14:01.615662 6527 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.835895 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.845599 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.845647 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.845658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.845672 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.845683 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:23Z","lastTransitionTime":"2025-12-03T17:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.852121 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455eaaf8-262f-4da5-9040-76283bae7a46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80c7d63868c2a5ff22d609f1c3bb94a939eca71443e7b274c77219c4b641e12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.869771 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.886954 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.901934 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.920212 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.945795 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.948259 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.948322 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.948345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.948377 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.948401 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:23Z","lastTransitionTime":"2025-12-03T17:14:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.964225 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.981143 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:09Z\\\",\\\"message\\\":\\\"2025-12-03T17:13:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9\\\\n2025-12-03T17:13:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9 to /host/opt/cni/bin/\\\\n2025-12-03T17:13:24Z [verbose] multus-daemon started\\\\n2025-12-03T17:13:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T17:14:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:14:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:23 crc kubenswrapper[4787]: I1203 17:14:23.994697 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:23Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.014636 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.028486 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.043846 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.051274 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.051328 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.051341 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.051360 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.051372 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:24Z","lastTransitionTime":"2025-12-03T17:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.062231 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.077425 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.092767 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:24Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.153123 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.153188 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.153199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.153217 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.153227 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:24Z","lastTransitionTime":"2025-12-03T17:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.256375 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.256429 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.256440 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.256459 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.256472 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:24Z","lastTransitionTime":"2025-12-03T17:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.358867 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.358909 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.358928 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.358950 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.358965 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:24Z","lastTransitionTime":"2025-12-03T17:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.461550 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.461610 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.461625 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.461649 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.461665 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:24Z","lastTransitionTime":"2025-12-03T17:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.564233 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.564290 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.564299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.564314 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.564327 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:24Z","lastTransitionTime":"2025-12-03T17:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.687648 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.687731 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.687757 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.687784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.687806 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:24Z","lastTransitionTime":"2025-12-03T17:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.765973 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.766054 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.766158 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:24 crc kubenswrapper[4787]: E1203 17:14:24.766447 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:24 crc kubenswrapper[4787]: E1203 17:14:24.766611 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:24 crc kubenswrapper[4787]: E1203 17:14:24.766835 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.790745 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.790803 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.790813 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.790836 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.790849 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:24Z","lastTransitionTime":"2025-12-03T17:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.893915 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.893987 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.894007 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.894076 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.894097 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:24Z","lastTransitionTime":"2025-12-03T17:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.996644 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.996705 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.996722 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.996750 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:24 crc kubenswrapper[4787]: I1203 17:14:24.996768 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:24Z","lastTransitionTime":"2025-12-03T17:14:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.099811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.099884 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.099905 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.099933 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.099954 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:25Z","lastTransitionTime":"2025-12-03T17:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.202651 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.202964 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.203152 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.203340 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.203534 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:25Z","lastTransitionTime":"2025-12-03T17:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.306803 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.306871 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.306888 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.306913 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.306933 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:25Z","lastTransitionTime":"2025-12-03T17:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.409917 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.410281 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.410374 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.410483 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.410576 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:25Z","lastTransitionTime":"2025-12-03T17:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.514083 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.514129 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.514142 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.514158 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.514170 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:25Z","lastTransitionTime":"2025-12-03T17:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.617667 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.617712 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.617729 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.617755 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.617772 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:25Z","lastTransitionTime":"2025-12-03T17:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.721856 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.722071 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.722163 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.722253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.722328 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:25Z","lastTransitionTime":"2025-12-03T17:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.768835 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:25 crc kubenswrapper[4787]: E1203 17:14:25.768935 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.769604 4787 scope.go:117] "RemoveContainer" containerID="91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.825313 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.825351 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.825359 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.825373 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.825384 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:25Z","lastTransitionTime":"2025-12-03T17:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.929182 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.930492 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.930687 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.930861 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:25 crc kubenswrapper[4787]: I1203 17:14:25.931160 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:25Z","lastTransitionTime":"2025-12-03T17:14:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.033675 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.033729 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.033756 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.033780 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.033797 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:26Z","lastTransitionTime":"2025-12-03T17:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.136945 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.136988 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.137001 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.137038 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.137053 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:26Z","lastTransitionTime":"2025-12-03T17:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.239648 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.239700 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.239709 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.239727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.239740 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:26Z","lastTransitionTime":"2025-12-03T17:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.284116 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/2.log" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.287084 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6"} Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.287672 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.300622 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.317667 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.342437 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.342468 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.342480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.342495 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.342504 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:26Z","lastTransitionTime":"2025-12-03T17:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.343054 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.374456 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:01Z\\\",\\\"message\\\":\\\"5-0021ba7e4488}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:14:01.614914 6527 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 17:14:01.614935 6527 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.182\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9099, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 17:14:01.614955 6527 services_controller.go:444] Built service openshift-cluster-version/cluster-version-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.614968 6527 services_controller.go:445] Built service openshift-cluster-version/cluster-version-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.615002 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:14:01.615662 6527 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.387209 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.396822 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455eaaf8-262f-4da5-9040-76283bae7a46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80c7d63868c2a5ff22d609f1c3bb94a939eca71443e7b274c77219c4b641e12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.408000 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.420304 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.433398 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:09Z\\\",\\\"message\\\":\\\"2025-12-03T17:13:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9\\\\n2025-12-03T17:13:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9 to /host/opt/cni/bin/\\\\n2025-12-03T17:13:24Z [verbose] multus-daemon started\\\\n2025-12-03T17:13:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T17:14:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:14:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.444299 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.444344 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.444357 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.448224 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.444374 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.452146 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:26Z","lastTransitionTime":"2025-12-03T17:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.470489 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.489862 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.512656 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.527714 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.541840 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.552556 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.554107 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.554177 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.554195 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.554222 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.554240 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:26Z","lastTransitionTime":"2025-12-03T17:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.569199 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.583573 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.598992 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.657296 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.657355 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.657365 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.657385 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.657397 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:26Z","lastTransitionTime":"2025-12-03T17:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.660762 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.660976 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.661082 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:30.660995624 +0000 UTC m=+147.478466603 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.661189 4787 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.661231 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.661273 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:15:30.661252181 +0000 UTC m=+147.478723130 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.661396 4787 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.661483 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-03 17:15:30.661467927 +0000 UTC m=+147.478939106 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.760588 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.760650 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.760672 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.760700 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.760716 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:26Z","lastTransitionTime":"2025-12-03T17:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.761788 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.761847 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.762169 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.762225 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.762247 4787 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.762349 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-03 17:15:30.76231349 +0000 UTC m=+147.579784679 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.762872 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.763128 4787 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.763326 4787 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.763630 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-03 17:15:30.763588324 +0000 UTC m=+147.581059453 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.764964 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.765122 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.765243 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.765425 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.765726 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:26 crc kubenswrapper[4787]: E1203 17:14:26.766096 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.864677 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.865608 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.865775 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.865918 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.866077 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:26Z","lastTransitionTime":"2025-12-03T17:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.968714 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.968784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.968794 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.968811 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:26 crc kubenswrapper[4787]: I1203 17:14:26.968823 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:26Z","lastTransitionTime":"2025-12-03T17:14:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.071784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.071841 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.071858 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.071882 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.071899 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.112558 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.112602 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.112614 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.112631 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.112643 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: E1203 17:14:27.134807 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.141603 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.141669 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.141698 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.141733 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.141756 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: E1203 17:14:27.163262 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.167790 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.167854 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.167873 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.167899 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.167918 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: E1203 17:14:27.188184 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.193768 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.193832 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.193849 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.193867 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.193881 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: E1203 17:14:27.208945 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.213373 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.213446 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.213471 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.213502 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.213524 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: E1203 17:14:27.230621 4787 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"6d35e1b1-0d51-401a-9c96-4da3ec675199\\\",\\\"systemUUID\\\":\\\"f359188a-c3b6-4925-b0ef-b6c8ba0e4e26\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: E1203 17:14:27.230893 4787 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.232956 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.233047 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.233067 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.233093 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.233116 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.293636 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/3.log" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.294513 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/2.log" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.298707 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6" exitCode=1 Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.298843 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6"} Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.298981 4787 scope.go:117] "RemoveContainer" containerID="91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.300063 4787 scope.go:117] "RemoveContainer" containerID="d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6" Dec 03 17:14:27 crc kubenswrapper[4787]: E1203 17:14:27.300427 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.320267 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.336508 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.336972 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.337083 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.337577 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.337670 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.342810 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.362746 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.387336 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://91306f89065c6ac864cf23d2da850e472120896fb07dfd13d8f264ce15bf2657\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:01Z\\\",\\\"message\\\":\\\"5-0021ba7e4488}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1203 17:14:01.614914 6527 lb_config.go:1031] Cluster endpoints for openshift-cluster-version/cluster-version-operator for network=default are: map[]\\\\nI1203 17:14:01.614935 6527 services_controller.go:443] Built service openshift-cluster-version/cluster-version-operator LB cluster-wide configs for network=default: []services.lbConfig{services.lbConfig{vips:[]string{\\\\\\\"10.217.4.182\\\\\\\"}, protocol:\\\\\\\"TCP\\\\\\\", inport:9099, clusterEndpoints:services.lbEndpoints{Port:0, V4IPs:[]string(nil), V6IPs:[]string(nil)}, nodeEndpoints:map[string]services.lbEndpoints{}, externalTrafficLocal:false, internalTrafficLocal:false, hasNodePort:false}}\\\\nI1203 17:14:01.614955 6527 services_controller.go:444] Built service openshift-cluster-version/cluster-version-operator LB per-node configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.614968 6527 services_controller.go:445] Built service openshift-cluster-version/cluster-version-operator LB template configs for network=default: []services.lbConfig(nil)\\\\nI1203 17:14:01.615002 6527 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1203 17:14:01.615662 6527 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:00Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 17:14:26.739580 6942 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-xnl29\\\\nI1203 17:14:26.739660 6942 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-xnl29\\\\nF1203 17:14:26.739687 6942 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z]\\\\nI1203 17:14:26.739705 6942 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-xnl29 in node crc\\\\nI1203 17:14:26.739572 6942 default_network_controller.go:776] Recording success event \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.399361 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.413991 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455eaaf8-262f-4da5-9040-76283bae7a46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80c7d63868c2a5ff22d609f1c3bb94a939eca71443e7b274c77219c4b641e12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.433608 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.440773 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.440857 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.440877 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.440908 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.440930 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.449365 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.464812 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:09Z\\\",\\\"message\\\":\\\"2025-12-03T17:13:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9\\\\n2025-12-03T17:13:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9 to /host/opt/cni/bin/\\\\n2025-12-03T17:13:24Z [verbose] multus-daemon started\\\\n2025-12-03T17:13:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T17:14:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:14:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.483141 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.505910 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.522435 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.538679 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.545104 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.545179 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.545199 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.545227 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.545246 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.555838 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.571144 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.582245 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.597314 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.614217 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.627570 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:27Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.647750 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.647803 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.647815 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.647835 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.647848 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.750698 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.750794 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.750814 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.750838 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.750856 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.766226 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:27 crc kubenswrapper[4787]: E1203 17:14:27.766418 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.853994 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.854045 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.854073 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.854088 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.854097 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.957941 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.958000 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.958030 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.958052 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:27 crc kubenswrapper[4787]: I1203 17:14:27.958066 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:27Z","lastTransitionTime":"2025-12-03T17:14:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.062091 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.062140 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.062156 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.062179 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.062196 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:28Z","lastTransitionTime":"2025-12-03T17:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.165127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.165252 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.165273 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.165300 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.165317 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:28Z","lastTransitionTime":"2025-12-03T17:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.268829 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.269073 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.269095 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.269119 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.269174 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:28Z","lastTransitionTime":"2025-12-03T17:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.306656 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/3.log" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.312331 4787 scope.go:117] "RemoveContainer" containerID="d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6" Dec 03 17:14:28 crc kubenswrapper[4787]: E1203 17:14:28.313433 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.328731 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.346396 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.371355 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.371417 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.371434 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.371460 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.371478 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:28Z","lastTransitionTime":"2025-12-03T17:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.377779 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 17:14:26.739580 6942 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-xnl29\\\\nI1203 17:14:26.739660 6942 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-xnl29\\\\nF1203 17:14:26.739687 6942 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z]\\\\nI1203 17:14:26.739705 6942 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-xnl29 in node crc\\\\nI1203 17:14:26.739572 6942 default_network_controller.go:776] Recording success event \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.397112 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.412926 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455eaaf8-262f-4da5-9040-76283bae7a46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80c7d63868c2a5ff22d609f1c3bb94a939eca71443e7b274c77219c4b641e12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.432227 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.451260 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.465244 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.474346 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.474396 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.474409 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.474427 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.474439 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:28Z","lastTransitionTime":"2025-12-03T17:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.483404 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.516796 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.535325 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.552424 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:09Z\\\",\\\"message\\\":\\\"2025-12-03T17:13:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9\\\\n2025-12-03T17:13:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9 to /host/opt/cni/bin/\\\\n2025-12-03T17:13:24Z [verbose] multus-daemon started\\\\n2025-12-03T17:13:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T17:14:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:14:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.565454 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.577910 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.577974 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.577993 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.578049 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.578072 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:28Z","lastTransitionTime":"2025-12-03T17:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.588688 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.614504 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.632334 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.654695 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.676490 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.681014 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.681072 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.681083 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.681099 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.681108 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:28Z","lastTransitionTime":"2025-12-03T17:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.698066 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:28Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.765303 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.765388 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:28 crc kubenswrapper[4787]: E1203 17:14:28.765468 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:28 crc kubenswrapper[4787]: E1203 17:14:28.765619 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.765745 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:28 crc kubenswrapper[4787]: E1203 17:14:28.765843 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.783994 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.784075 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.784092 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.784117 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.784132 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:28Z","lastTransitionTime":"2025-12-03T17:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.887183 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.887250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.887273 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.887300 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.887322 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:28Z","lastTransitionTime":"2025-12-03T17:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.990764 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.990821 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.990837 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.990863 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:28 crc kubenswrapper[4787]: I1203 17:14:28.990903 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:28Z","lastTransitionTime":"2025-12-03T17:14:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.094234 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.094303 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.094316 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.094345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.094359 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:29Z","lastTransitionTime":"2025-12-03T17:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.198002 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.198114 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.198135 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.198161 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.198179 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:29Z","lastTransitionTime":"2025-12-03T17:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.300795 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.300840 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.300854 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.300871 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.300882 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:29Z","lastTransitionTime":"2025-12-03T17:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.403287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.403346 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.403358 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.403378 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.403394 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:29Z","lastTransitionTime":"2025-12-03T17:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.506671 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.506770 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.506791 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.506809 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.506821 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:29Z","lastTransitionTime":"2025-12-03T17:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.610101 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.610175 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.610195 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.610220 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.610236 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:29Z","lastTransitionTime":"2025-12-03T17:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.714318 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.714831 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.714850 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.714882 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.714903 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:29Z","lastTransitionTime":"2025-12-03T17:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.765929 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:29 crc kubenswrapper[4787]: E1203 17:14:29.766197 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.818566 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.818624 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.818641 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.818666 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.818684 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:29Z","lastTransitionTime":"2025-12-03T17:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.923041 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.923077 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.923088 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.923105 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:29 crc kubenswrapper[4787]: I1203 17:14:29.923115 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:29Z","lastTransitionTime":"2025-12-03T17:14:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.025833 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.025884 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.025898 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.025916 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.025929 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:30Z","lastTransitionTime":"2025-12-03T17:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.129656 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.129737 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.129754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.129779 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.129791 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:30Z","lastTransitionTime":"2025-12-03T17:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.232963 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.233044 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.233059 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.233076 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.233087 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:30Z","lastTransitionTime":"2025-12-03T17:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.336789 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.336844 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.336859 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.337077 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.337093 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:30Z","lastTransitionTime":"2025-12-03T17:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.439483 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.439559 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.439574 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.439600 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.439617 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:30Z","lastTransitionTime":"2025-12-03T17:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.541797 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.541837 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.541846 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.541858 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.541867 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:30Z","lastTransitionTime":"2025-12-03T17:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.644892 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.644973 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.644990 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.645012 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.645040 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:30Z","lastTransitionTime":"2025-12-03T17:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.748985 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.749076 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.749089 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.749113 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.749128 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:30Z","lastTransitionTime":"2025-12-03T17:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.765203 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.765240 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.765390 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:30 crc kubenswrapper[4787]: E1203 17:14:30.765523 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:30 crc kubenswrapper[4787]: E1203 17:14:30.765589 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:30 crc kubenswrapper[4787]: E1203 17:14:30.765724 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.851588 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.851650 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.851663 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.851681 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.851693 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:30Z","lastTransitionTime":"2025-12-03T17:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.954098 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.954159 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.954171 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.954188 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:30 crc kubenswrapper[4787]: I1203 17:14:30.954202 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:30Z","lastTransitionTime":"2025-12-03T17:14:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.057047 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.057114 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.057125 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.057144 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.057155 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:31Z","lastTransitionTime":"2025-12-03T17:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.159295 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.159364 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.159382 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.159404 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.159425 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:31Z","lastTransitionTime":"2025-12-03T17:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.262659 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.262792 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.262812 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.262836 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.262854 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:31Z","lastTransitionTime":"2025-12-03T17:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.366560 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.366633 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.366660 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.366697 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.366725 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:31Z","lastTransitionTime":"2025-12-03T17:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.470003 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.470090 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.470107 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.470130 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.470150 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:31Z","lastTransitionTime":"2025-12-03T17:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.573138 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.573210 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.573228 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.573250 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.573267 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:31Z","lastTransitionTime":"2025-12-03T17:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.675922 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.675987 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.675998 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.676038 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.676050 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:31Z","lastTransitionTime":"2025-12-03T17:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.765504 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:31 crc kubenswrapper[4787]: E1203 17:14:31.765678 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.778645 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.778702 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.778713 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.778727 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.778763 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:31Z","lastTransitionTime":"2025-12-03T17:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.881842 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.881896 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.881910 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.881930 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.881942 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:31Z","lastTransitionTime":"2025-12-03T17:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.985561 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.985627 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.985645 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.985689 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:31 crc kubenswrapper[4787]: I1203 17:14:31.985708 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:31Z","lastTransitionTime":"2025-12-03T17:14:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.089271 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.089347 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.089378 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.089406 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.089422 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:32Z","lastTransitionTime":"2025-12-03T17:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.193391 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.193418 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.193429 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.193448 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.193458 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:32Z","lastTransitionTime":"2025-12-03T17:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.296689 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.296759 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.296792 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.296821 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.296841 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:32Z","lastTransitionTime":"2025-12-03T17:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.399849 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.399895 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.399911 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.399934 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.399950 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:32Z","lastTransitionTime":"2025-12-03T17:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.503084 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.503127 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.503139 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.503155 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.503168 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:32Z","lastTransitionTime":"2025-12-03T17:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.604865 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.604935 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.604952 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.604977 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.604994 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:32Z","lastTransitionTime":"2025-12-03T17:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.707974 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.708073 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.708090 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.708113 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.708128 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:32Z","lastTransitionTime":"2025-12-03T17:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.766684 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:32 crc kubenswrapper[4787]: E1203 17:14:32.766775 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.766978 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:32 crc kubenswrapper[4787]: E1203 17:14:32.767061 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.767170 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:32 crc kubenswrapper[4787]: E1203 17:14:32.767214 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.811223 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.811305 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.811325 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.811358 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.811378 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:32Z","lastTransitionTime":"2025-12-03T17:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.913700 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.913756 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.913766 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.913786 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:32 crc kubenswrapper[4787]: I1203 17:14:32.913799 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:32Z","lastTransitionTime":"2025-12-03T17:14:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.017253 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.017334 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.017359 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.017399 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.017423 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:33Z","lastTransitionTime":"2025-12-03T17:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.120470 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.120510 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.120518 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.120532 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.120541 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:33Z","lastTransitionTime":"2025-12-03T17:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.223308 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.223344 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.223352 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.223365 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.223373 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:33Z","lastTransitionTime":"2025-12-03T17:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.326324 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.326408 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.326422 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.326438 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.326463 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:33Z","lastTransitionTime":"2025-12-03T17:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.429623 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.429669 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.429679 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.429696 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.429710 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:33Z","lastTransitionTime":"2025-12-03T17:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.532489 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.532543 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.532555 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.532571 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.532584 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:33Z","lastTransitionTime":"2025-12-03T17:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.635708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.635757 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.635774 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.635796 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.635813 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:33Z","lastTransitionTime":"2025-12-03T17:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.738717 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.738784 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.738801 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.738824 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.738842 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:33Z","lastTransitionTime":"2025-12-03T17:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.765623 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:33 crc kubenswrapper[4787]: E1203 17:14:33.765766 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.785794 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.803410 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b6597ac6-3ab2-4d2f-b38e-896795a7773d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78fca48b90d8cde16f41d8b727aa016c51ced163078399f317c15f076bc61b8c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-df4v9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-t882k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.820934 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef833e74-df41-4fe8-bec4-b5cf3541aa99\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f150340145af57b0b49f1d4c676b952463729c4becaf6b197482d4ac91bb6ddf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d6b979d7f3da0a18ba14184b993dcb7234afc80a0e52a9d4ef5adbfd19656732\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-2x886\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:34Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-6zkg7\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.838265 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"455eaaf8-262f-4da5-9040-76283bae7a46\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://80c7d63868c2a5ff22d609f1c3bb94a939eca71443e7b274c77219c4b641e12d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://b8047dbb4f455b8194b91681533d0e989c52ca95d30b62653d0cfae7d1f0fb97\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.842655 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.842783 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.842874 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.842969 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.843088 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:33Z","lastTransitionTime":"2025-12-03T17:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.878287 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ebdb4d20-1021-46ae-af76-e4bbb89cd048\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ef80b75abb3b3dad8e366a958548a8242a8880f3009bce603aa0541ba74c73bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c315505851b0ec3217671b7a79e9ae1f2ea26e52f5457b52069855d3c5657cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://3ebdf02f0bb426e9bda2b0a6f059806ffe92dcf6d39ad0a4a8202a79f80d638e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://24a42531d26a7131bdee1e7bf9ec01348f9bdd24369b99381f539856603c9d66\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b9b00dab289ad3e47349caf051aec60af4fe68a0588ecd8fae825866274c02b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aa810cb484c874b132c8f8398f0f8aada188868cd44673e30d725ecd29a6d6c1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://31fd443c2d489422b78d8f4014a82c6ba8e6146b4bf2580ba85bb531da1bcaeb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://99bf30ca4064cea108d13eb86e5b4b9d8a266ed2b1c6fe9962b375f10aa8891b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.899222 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f0042ff64d56e0a47c0a72c1e52a047f983c0d24c70ca0e1dadb3aa343c1cc1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.922438 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-xcz4f" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"bd5617c0-aa4f-4f21-a131-15af831725ec\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:14:11Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:09Z\\\",\\\"message\\\":\\\"2025-12-03T17:13:24+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9\\\\n2025-12-03T17:13:24+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_a7a11e18-b156-46e0-b2b6-3f343ca7f3c9 to /host/opt/cni/bin/\\\\n2025-12-03T17:13:24Z [verbose] multus-daemon started\\\\n2025-12-03T17:13:24Z [verbose] Readiness Indicator file check\\\\n2025-12-03T17:14:09Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:14:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-tdlvg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-xcz4f\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.942249 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3239dbc7-323e-4402-90e8-615da2c72ecb\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f884b467df0ff6c05a2bc93116050768cd4ba1aad5890fa7ad03ec12eeb9bcba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1aa32749f08b4bb9ed00fce6bdc219e80dc24bb3a094d7242b17fb96ca130c5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26cbbcf231406388072806cbd37f5369ce26605f41944592f3f1f585981a2926\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c35acffd291d368294b25f0156431c0285da97e05a4130357dc7e751fcdd0ccd\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.947088 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.947156 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.947170 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.947194 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.947216 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:33Z","lastTransitionTime":"2025-12-03T17:14:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.962029 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7a92940f-2d9e-466f-8cfd-5a3cf91740f6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1203 17:13:16.241857 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1203 17:13:16.243526 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3723381787/tls.crt::/tmp/serving-cert-3723381787/tls.key\\\\\\\"\\\\nI1203 17:13:21.932750 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1203 17:13:21.937224 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1203 17:13:21.937245 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1203 17:13:21.937267 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1203 17:13:21.937284 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1203 17:13:21.944492 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1203 17:13:21.944515 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944519 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1203 17:13:21.944524 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1203 17:13:21.944527 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1203 17:13:21.944530 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1203 17:13:21.944533 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1203 17:13:21.944737 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1203 17:13:21.951034 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.978227 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:33 crc kubenswrapper[4787]: I1203 17:14:33.993546 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:33Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.006942 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-5rllg" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1f5d3d8c-cdec-4ed3-bb53-0a5012751616\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a596b76a0008e6d500448b0cce22a81248f05623f936bc6be36e82adcfe3c2bf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-l2dfr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-5rllg\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.030247 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"34f0f91e-2de6-4e13-974c-917d871c3a43\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3c3fec868b78ddaf6f247bcac5ee781565d9f08ad92f368b7f5560fae36bac09\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:31Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cef26f638b8d64243f415b0182e4204313c3ad6497a31ca351a3533584aa3fe\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ddc7f8a370ab5043c3366e84a56693133e7be56abd4f5275f977e525dd4ae820\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://afc1747f80631b7a9fc5e20095fac1bafd71b52f343c5a3262ea9d71b17cacd8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:26Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:26Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://14acb18eb808581abb6a7eebd20f85f854660af8c332948acaf6052b7b3dc4b5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:27Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8a490fa5b0b9abc65ad27470e07125b42a22ccb79a2fdc982d583fd82cbbc444\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:29Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:28Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bcb169ff631f99dfd1ec820679b9cc903b76f86e7573340d106975958fde3dd3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:30Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rrl8n\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vj6t6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.043973 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"50c2569b-2a14-4112-82e4-afc683aa36a7\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:36Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-wbgkj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:36Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-m9tr6\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.049193 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.049257 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.049268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.049287 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.049300 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:34Z","lastTransitionTime":"2025-12-03T17:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.061115 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"88257c49-5dca-40f4-b6f6-6a95e85eda2c\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8b0113f42891f5f9cd08eb6d394e5ed672b11bf0b9d50ff302d0e1ae858da462\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f1b6b693fe6bfcec3946986e08309e0320950166375182d834c28999b209afbc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://bab53be2610030422f169b95468039f5cde7bc0b2656a5b37c2af28d44429be2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:03Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.073681 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:28Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c225328d14153c0b9873a5846cf8d2cb09135493e229db46e6c7a5f2ea55539b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.089956 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c3f5a989-6421-4bc7-a516-b37c825d8bbc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:22Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-03T17:14:27Z\\\",\\\"message\\\":\\\"Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1203 17:14:26.739580 6942 obj_retry.go:303] Retry object setup: *v1.Pod openshift-image-registry/node-ca-xnl29\\\\nI1203 17:14:26.739660 6942 obj_retry.go:365] Adding new object: *v1.Pod openshift-image-registry/node-ca-xnl29\\\\nF1203 17:14:26.739687 6942 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:26Z is after 2025-08-24T17:21:41Z]\\\\nI1203 17:14:26.739705 6942 ovn.go:134] Ensuring zone local for Pod openshift-image-registry/node-ca-xnl29 in node crc\\\\nI1203 17:14:26.739572 6942 default_network_controller.go:776] Recording success event \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-03T17:14:25Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:27Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5k4bd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:22Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-bx8nf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.105079 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-xnl29" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9c61ffb2-8f34-4a8b-9e65-89da492c3382\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:25Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://23c4c03e5603db54aa6f0eb844e8c9fe58a43d8d0a4794196f6f7502264b98ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:25Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-j5xbx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-03T17:13:24Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-xnl29\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.121636 4787 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-03T17:13:23Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8d916bd98e3dfe642ae1e07a7c1ff175f6598f1fc58c47750e6911fbdcb52e47\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2ca5309d6c87b025e2b0d7e3175352cdf2d228fc6580d708a060a15db8fc714f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-03T17:13:23Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-03T17:14:34Z is after 2025-08-24T17:21:41Z" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.152404 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.152434 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.152460 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.152475 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.152485 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:34Z","lastTransitionTime":"2025-12-03T17:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.255795 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.255822 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.255831 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.255843 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.255851 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:34Z","lastTransitionTime":"2025-12-03T17:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.359056 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.359096 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.359110 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.359131 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.359145 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:34Z","lastTransitionTime":"2025-12-03T17:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.461174 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.461212 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.461223 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.461239 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.461250 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:34Z","lastTransitionTime":"2025-12-03T17:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.563620 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.563658 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.563668 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.563684 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.563695 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:34Z","lastTransitionTime":"2025-12-03T17:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.666480 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.666527 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.666540 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.666556 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.666568 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:34Z","lastTransitionTime":"2025-12-03T17:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.765366 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:34 crc kubenswrapper[4787]: E1203 17:14:34.765732 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.765740 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:34 crc kubenswrapper[4787]: E1203 17:14:34.765968 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.765757 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:34 crc kubenswrapper[4787]: E1203 17:14:34.766256 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.768084 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.768118 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.768129 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.768142 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.768151 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:34Z","lastTransitionTime":"2025-12-03T17:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.870798 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.870859 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.870875 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.870901 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.870920 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:34Z","lastTransitionTime":"2025-12-03T17:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.973637 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.973691 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.973708 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.973734 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:34 crc kubenswrapper[4787]: I1203 17:14:34.973752 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:34Z","lastTransitionTime":"2025-12-03T17:14:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.076277 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.076326 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.076339 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.076356 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.076371 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:35Z","lastTransitionTime":"2025-12-03T17:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.179200 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.179268 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.179288 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.179312 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.179329 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:35Z","lastTransitionTime":"2025-12-03T17:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.281407 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.281513 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.281532 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.281556 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.281576 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:35Z","lastTransitionTime":"2025-12-03T17:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.384310 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.384355 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.384367 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.384384 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.384397 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:35Z","lastTransitionTime":"2025-12-03T17:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.487692 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.487740 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.487754 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.487772 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.487786 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:35Z","lastTransitionTime":"2025-12-03T17:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.590823 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.590871 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.590883 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.590898 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.590908 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:35Z","lastTransitionTime":"2025-12-03T17:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.693894 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.694172 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.694190 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.694213 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.694227 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:35Z","lastTransitionTime":"2025-12-03T17:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.765596 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:35 crc kubenswrapper[4787]: E1203 17:14:35.765796 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.797094 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.797141 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.797155 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.797171 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.797181 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:35Z","lastTransitionTime":"2025-12-03T17:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.903705 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.903779 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.903804 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.903834 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:35 crc kubenswrapper[4787]: I1203 17:14:35.903856 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:35Z","lastTransitionTime":"2025-12-03T17:14:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.006948 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.007141 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.007174 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.007195 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.007208 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:36Z","lastTransitionTime":"2025-12-03T17:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.110351 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.110416 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.110431 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.110456 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.110473 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:36Z","lastTransitionTime":"2025-12-03T17:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.213793 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.213855 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.213872 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.213894 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.213910 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:36Z","lastTransitionTime":"2025-12-03T17:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.316242 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.316309 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.316318 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.316337 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.316353 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:36Z","lastTransitionTime":"2025-12-03T17:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.419302 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.419362 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.419379 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.419401 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.419418 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:36Z","lastTransitionTime":"2025-12-03T17:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.522207 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.522266 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.522282 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.522305 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.522323 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:36Z","lastTransitionTime":"2025-12-03T17:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.625252 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.625345 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.625370 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.625399 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.625421 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:36Z","lastTransitionTime":"2025-12-03T17:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.728580 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.728633 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.728649 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.728672 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.728688 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:36Z","lastTransitionTime":"2025-12-03T17:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.765843 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:36 crc kubenswrapper[4787]: E1203 17:14:36.766226 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.766273 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.766405 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:36 crc kubenswrapper[4787]: E1203 17:14:36.766485 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:36 crc kubenswrapper[4787]: E1203 17:14:36.766588 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.831401 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.831466 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.831485 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.831511 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.831530 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:36Z","lastTransitionTime":"2025-12-03T17:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.933639 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.933695 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.933711 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.933733 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:36 crc kubenswrapper[4787]: I1203 17:14:36.933751 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:36Z","lastTransitionTime":"2025-12-03T17:14:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.037096 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.037170 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.037183 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.037206 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.037221 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:37Z","lastTransitionTime":"2025-12-03T17:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.140847 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.140903 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.140920 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.140944 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.140962 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:37Z","lastTransitionTime":"2025-12-03T17:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.243998 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.244099 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.244121 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.244152 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.244175 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:37Z","lastTransitionTime":"2025-12-03T17:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.345876 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.345948 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.345966 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.345989 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.346078 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:37Z","lastTransitionTime":"2025-12-03T17:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.422749 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.422830 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.422854 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.422881 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.422905 4787 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-03T17:14:37Z","lastTransitionTime":"2025-12-03T17:14:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.476754 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr"] Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.477338 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.480462 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 17:14:37 crc kubenswrapper[4787]: W1203 17:14:37.480505 4787 reflector.go:561] object-"openshift-cluster-version"/"default-dockercfg-gxtc4": failed to list *v1.Secret: secrets "default-dockercfg-gxtc4" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-version": no relationship found between node 'crc' and this object Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.480524 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 17:14:37 crc kubenswrapper[4787]: E1203 17:14:37.480557 4787 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-version\"/\"default-dockercfg-gxtc4\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"default-dockercfg-gxtc4\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-version\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.480795 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.521614 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=20.521594397 podStartE2EDuration="20.521594397s" podCreationTimestamp="2025-12-03 17:14:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:37.504240927 +0000 UTC m=+94.321711916" watchObservedRunningTime="2025-12-03 17:14:37.521594397 +0000 UTC m=+94.339065346" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.562729 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podStartSLOduration=75.562709627 podStartE2EDuration="1m15.562709627s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:37.541414712 +0000 UTC m=+94.358885681" watchObservedRunningTime="2025-12-03 17:14:37.562709627 +0000 UTC m=+94.380180586" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.583879 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-6zkg7" podStartSLOduration=75.583851747 podStartE2EDuration="1m15.583851747s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:37.56320775 +0000 UTC m=+94.380678709" watchObservedRunningTime="2025-12-03 17:14:37.583851747 +0000 UTC m=+94.401322726" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.587343 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.587398 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.587423 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.587445 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.587492 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.625624 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=47.625605884 podStartE2EDuration="47.625605884s" podCreationTimestamp="2025-12-03 17:13:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:37.601163966 +0000 UTC m=+94.418634925" watchObservedRunningTime="2025-12-03 17:14:37.625605884 +0000 UTC m=+94.443076843" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.640567 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=73.64054802 podStartE2EDuration="1m13.64054802s" podCreationTimestamp="2025-12-03 17:13:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:37.626141978 +0000 UTC m=+94.443612957" watchObservedRunningTime="2025-12-03 17:14:37.64054802 +0000 UTC m=+94.458018979" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.679908 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=76.679891192 podStartE2EDuration="1m16.679891192s" podCreationTimestamp="2025-12-03 17:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:37.678922257 +0000 UTC m=+94.496393226" watchObservedRunningTime="2025-12-03 17:14:37.679891192 +0000 UTC m=+94.497362151" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.680043 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-xcz4f" podStartSLOduration=75.680040196 podStartE2EDuration="1m15.680040196s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:37.66660166 +0000 UTC m=+94.484072629" watchObservedRunningTime="2025-12-03 17:14:37.680040196 +0000 UTC m=+94.497511155" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.688690 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.688760 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.688782 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.688799 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.688815 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.688863 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.688897 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.689660 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-service-ca\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.694975 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.711786 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60290afe-5158-4e23-aa2d-1e90c5dfcdaf-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-b9rxr\" (UID: \"60290afe-5158-4e23-aa2d-1e90c5dfcdaf\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.744684 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=75.744663869 podStartE2EDuration="1m15.744663869s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:37.723004745 +0000 UTC m=+94.540475714" watchObservedRunningTime="2025-12-03 17:14:37.744663869 +0000 UTC m=+94.562134818" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.765784 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:37 crc kubenswrapper[4787]: E1203 17:14:37.765903 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.794623 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-5rllg" podStartSLOduration=76.794606583 podStartE2EDuration="1m16.794606583s" podCreationTimestamp="2025-12-03 17:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:37.778819765 +0000 UTC m=+94.596290724" watchObservedRunningTime="2025-12-03 17:14:37.794606583 +0000 UTC m=+94.612077542" Dec 03 17:14:37 crc kubenswrapper[4787]: I1203 17:14:37.794709 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-vj6t6" podStartSLOduration=75.794706246 podStartE2EDuration="1m15.794706246s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:37.794368727 +0000 UTC m=+94.611839696" watchObservedRunningTime="2025-12-03 17:14:37.794706246 +0000 UTC m=+94.612177205" Dec 03 17:14:38 crc kubenswrapper[4787]: I1203 17:14:38.765454 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:38 crc kubenswrapper[4787]: I1203 17:14:38.765559 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:38 crc kubenswrapper[4787]: I1203 17:14:38.765595 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:38 crc kubenswrapper[4787]: E1203 17:14:38.765909 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:38 crc kubenswrapper[4787]: E1203 17:14:38.766071 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:38 crc kubenswrapper[4787]: E1203 17:14:38.766172 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:38 crc kubenswrapper[4787]: I1203 17:14:38.802996 4787 kubelet_pods.go:1007] "Unable to retrieve pull secret, the image pull may not succeed." pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" secret="" err="failed to sync secret cache: timed out waiting for the condition" Dec 03 17:14:38 crc kubenswrapper[4787]: I1203 17:14:38.803089 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" Dec 03 17:14:38 crc kubenswrapper[4787]: I1203 17:14:38.949042 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 17:14:39 crc kubenswrapper[4787]: I1203 17:14:39.348365 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" event={"ID":"60290afe-5158-4e23-aa2d-1e90c5dfcdaf","Type":"ContainerStarted","Data":"5f8447ab968a8c0197d8ce2461a839714d479ecb6714555ae81974d6a0b8d78c"} Dec 03 17:14:39 crc kubenswrapper[4787]: I1203 17:14:39.348440 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" event={"ID":"60290afe-5158-4e23-aa2d-1e90c5dfcdaf","Type":"ContainerStarted","Data":"1d5c49a01f255ab3070558404696c0656b2b95d9b767f86768beb100f5cfdcd6"} Dec 03 17:14:39 crc kubenswrapper[4787]: I1203 17:14:39.362094 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-b9rxr" podStartSLOduration=77.362074989 podStartE2EDuration="1m17.362074989s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:39.361216866 +0000 UTC m=+96.178687825" watchObservedRunningTime="2025-12-03 17:14:39.362074989 +0000 UTC m=+96.179545958" Dec 03 17:14:39 crc kubenswrapper[4787]: I1203 17:14:39.362505 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-xnl29" podStartSLOduration=78.36249801 podStartE2EDuration="1m18.36249801s" podCreationTimestamp="2025-12-03 17:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:14:37.878917848 +0000 UTC m=+94.696388817" watchObservedRunningTime="2025-12-03 17:14:39.36249801 +0000 UTC m=+96.179968979" Dec 03 17:14:39 crc kubenswrapper[4787]: I1203 17:14:39.766473 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:39 crc kubenswrapper[4787]: E1203 17:14:39.766714 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:40 crc kubenswrapper[4787]: I1203 17:14:40.414506 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:40 crc kubenswrapper[4787]: E1203 17:14:40.414661 4787 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:14:40 crc kubenswrapper[4787]: E1203 17:14:40.414752 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs podName:50c2569b-2a14-4112-82e4-afc683aa36a7 nodeName:}" failed. No retries permitted until 2025-12-03 17:15:44.41473366 +0000 UTC m=+161.232204619 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs") pod "network-metrics-daemon-m9tr6" (UID: "50c2569b-2a14-4112-82e4-afc683aa36a7") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 03 17:14:40 crc kubenswrapper[4787]: I1203 17:14:40.765169 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:40 crc kubenswrapper[4787]: I1203 17:14:40.765301 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:40 crc kubenswrapper[4787]: E1203 17:14:40.765333 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:40 crc kubenswrapper[4787]: I1203 17:14:40.765387 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:40 crc kubenswrapper[4787]: E1203 17:14:40.765526 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:40 crc kubenswrapper[4787]: E1203 17:14:40.765687 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:41 crc kubenswrapper[4787]: I1203 17:14:41.765484 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:41 crc kubenswrapper[4787]: E1203 17:14:41.765680 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:42 crc kubenswrapper[4787]: I1203 17:14:42.765439 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:42 crc kubenswrapper[4787]: I1203 17:14:42.766204 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:42 crc kubenswrapper[4787]: I1203 17:14:42.765444 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:42 crc kubenswrapper[4787]: E1203 17:14:42.766339 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:42 crc kubenswrapper[4787]: E1203 17:14:42.766491 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:42 crc kubenswrapper[4787]: E1203 17:14:42.766560 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:43 crc kubenswrapper[4787]: I1203 17:14:43.765160 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:43 crc kubenswrapper[4787]: E1203 17:14:43.765945 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:43 crc kubenswrapper[4787]: I1203 17:14:43.767158 4787 scope.go:117] "RemoveContainer" containerID="d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6" Dec 03 17:14:43 crc kubenswrapper[4787]: E1203 17:14:43.767391 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" Dec 03 17:14:44 crc kubenswrapper[4787]: I1203 17:14:44.765923 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:44 crc kubenswrapper[4787]: I1203 17:14:44.765960 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:44 crc kubenswrapper[4787]: E1203 17:14:44.766106 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:44 crc kubenswrapper[4787]: I1203 17:14:44.766134 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:44 crc kubenswrapper[4787]: E1203 17:14:44.766243 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:44 crc kubenswrapper[4787]: E1203 17:14:44.766421 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:45 crc kubenswrapper[4787]: I1203 17:14:45.765622 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:45 crc kubenswrapper[4787]: E1203 17:14:45.766194 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:46 crc kubenswrapper[4787]: I1203 17:14:46.765736 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:46 crc kubenswrapper[4787]: E1203 17:14:46.766234 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:46 crc kubenswrapper[4787]: I1203 17:14:46.766514 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:46 crc kubenswrapper[4787]: I1203 17:14:46.766601 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:46 crc kubenswrapper[4787]: E1203 17:14:46.766758 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:46 crc kubenswrapper[4787]: E1203 17:14:46.767557 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:47 crc kubenswrapper[4787]: I1203 17:14:47.765133 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:47 crc kubenswrapper[4787]: E1203 17:14:47.765312 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:48 crc kubenswrapper[4787]: I1203 17:14:48.765378 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:48 crc kubenswrapper[4787]: I1203 17:14:48.765466 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:48 crc kubenswrapper[4787]: E1203 17:14:48.765591 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:48 crc kubenswrapper[4787]: I1203 17:14:48.765641 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:48 crc kubenswrapper[4787]: E1203 17:14:48.765708 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:48 crc kubenswrapper[4787]: E1203 17:14:48.765995 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:49 crc kubenswrapper[4787]: I1203 17:14:49.765912 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:49 crc kubenswrapper[4787]: E1203 17:14:49.766183 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:50 crc kubenswrapper[4787]: I1203 17:14:50.765246 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:50 crc kubenswrapper[4787]: I1203 17:14:50.765290 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:50 crc kubenswrapper[4787]: I1203 17:14:50.765367 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:50 crc kubenswrapper[4787]: E1203 17:14:50.765433 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:50 crc kubenswrapper[4787]: E1203 17:14:50.766324 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:50 crc kubenswrapper[4787]: E1203 17:14:50.766440 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:51 crc kubenswrapper[4787]: I1203 17:14:51.765885 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:51 crc kubenswrapper[4787]: E1203 17:14:51.766082 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:52 crc kubenswrapper[4787]: I1203 17:14:52.765320 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:52 crc kubenswrapper[4787]: I1203 17:14:52.765350 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:52 crc kubenswrapper[4787]: E1203 17:14:52.765427 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:52 crc kubenswrapper[4787]: I1203 17:14:52.765320 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:52 crc kubenswrapper[4787]: E1203 17:14:52.765509 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:52 crc kubenswrapper[4787]: E1203 17:14:52.765552 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:53 crc kubenswrapper[4787]: I1203 17:14:53.766091 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:53 crc kubenswrapper[4787]: E1203 17:14:53.766525 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:54 crc kubenswrapper[4787]: I1203 17:14:54.765061 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:54 crc kubenswrapper[4787]: I1203 17:14:54.765158 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:54 crc kubenswrapper[4787]: I1203 17:14:54.765255 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:54 crc kubenswrapper[4787]: E1203 17:14:54.765178 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:54 crc kubenswrapper[4787]: E1203 17:14:54.765460 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:54 crc kubenswrapper[4787]: E1203 17:14:54.765835 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:55 crc kubenswrapper[4787]: I1203 17:14:55.765839 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:55 crc kubenswrapper[4787]: E1203 17:14:55.766147 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:56 crc kubenswrapper[4787]: I1203 17:14:56.404008 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xcz4f_bd5617c0-aa4f-4f21-a131-15af831725ec/kube-multus/1.log" Dec 03 17:14:56 crc kubenswrapper[4787]: I1203 17:14:56.404702 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xcz4f_bd5617c0-aa4f-4f21-a131-15af831725ec/kube-multus/0.log" Dec 03 17:14:56 crc kubenswrapper[4787]: I1203 17:14:56.404736 4787 generic.go:334] "Generic (PLEG): container finished" podID="bd5617c0-aa4f-4f21-a131-15af831725ec" containerID="59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8" exitCode=1 Dec 03 17:14:56 crc kubenswrapper[4787]: I1203 17:14:56.404762 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xcz4f" event={"ID":"bd5617c0-aa4f-4f21-a131-15af831725ec","Type":"ContainerDied","Data":"59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8"} Dec 03 17:14:56 crc kubenswrapper[4787]: I1203 17:14:56.404799 4787 scope.go:117] "RemoveContainer" containerID="d13086fa6442ab2ea5d9f0f68e94eda19174ee0e9714702f2ea41bad29b7c24d" Dec 03 17:14:56 crc kubenswrapper[4787]: I1203 17:14:56.405336 4787 scope.go:117] "RemoveContainer" containerID="59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8" Dec 03 17:14:56 crc kubenswrapper[4787]: E1203 17:14:56.405735 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-xcz4f_openshift-multus(bd5617c0-aa4f-4f21-a131-15af831725ec)\"" pod="openshift-multus/multus-xcz4f" podUID="bd5617c0-aa4f-4f21-a131-15af831725ec" Dec 03 17:14:56 crc kubenswrapper[4787]: I1203 17:14:56.764911 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:56 crc kubenswrapper[4787]: I1203 17:14:56.764983 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:56 crc kubenswrapper[4787]: E1203 17:14:56.765068 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:56 crc kubenswrapper[4787]: E1203 17:14:56.765158 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:56 crc kubenswrapper[4787]: I1203 17:14:56.764940 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:56 crc kubenswrapper[4787]: E1203 17:14:56.765361 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:57 crc kubenswrapper[4787]: I1203 17:14:57.415183 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xcz4f_bd5617c0-aa4f-4f21-a131-15af831725ec/kube-multus/1.log" Dec 03 17:14:57 crc kubenswrapper[4787]: I1203 17:14:57.765492 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:57 crc kubenswrapper[4787]: E1203 17:14:57.765919 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:14:57 crc kubenswrapper[4787]: I1203 17:14:57.766760 4787 scope.go:117] "RemoveContainer" containerID="d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6" Dec 03 17:14:57 crc kubenswrapper[4787]: E1203 17:14:57.766910 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-bx8nf_openshift-ovn-kubernetes(c3f5a989-6421-4bc7-a516-b37c825d8bbc)\"" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" Dec 03 17:14:58 crc kubenswrapper[4787]: I1203 17:14:58.766321 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:14:58 crc kubenswrapper[4787]: E1203 17:14:58.767184 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:14:58 crc kubenswrapper[4787]: I1203 17:14:58.766506 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:14:58 crc kubenswrapper[4787]: I1203 17:14:58.766463 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:14:58 crc kubenswrapper[4787]: E1203 17:14:58.767271 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:14:58 crc kubenswrapper[4787]: E1203 17:14:58.767535 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:14:59 crc kubenswrapper[4787]: I1203 17:14:59.766220 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:14:59 crc kubenswrapper[4787]: E1203 17:14:59.766459 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:15:00 crc kubenswrapper[4787]: I1203 17:15:00.765069 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:00 crc kubenswrapper[4787]: I1203 17:15:00.765143 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:00 crc kubenswrapper[4787]: E1203 17:15:00.765203 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:15:00 crc kubenswrapper[4787]: E1203 17:15:00.765366 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:15:00 crc kubenswrapper[4787]: I1203 17:15:00.765143 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:00 crc kubenswrapper[4787]: E1203 17:15:00.765497 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:15:01 crc kubenswrapper[4787]: I1203 17:15:01.765441 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:01 crc kubenswrapper[4787]: E1203 17:15:01.765578 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:15:02 crc kubenswrapper[4787]: I1203 17:15:02.765847 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:02 crc kubenswrapper[4787]: I1203 17:15:02.765911 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:02 crc kubenswrapper[4787]: I1203 17:15:02.765882 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:02 crc kubenswrapper[4787]: E1203 17:15:02.766054 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:15:02 crc kubenswrapper[4787]: E1203 17:15:02.766322 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:15:02 crc kubenswrapper[4787]: E1203 17:15:02.766444 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:15:03 crc kubenswrapper[4787]: E1203 17:15:03.761088 4787 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 03 17:15:03 crc kubenswrapper[4787]: I1203 17:15:03.765118 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:03 crc kubenswrapper[4787]: E1203 17:15:03.766401 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:15:03 crc kubenswrapper[4787]: E1203 17:15:03.846833 4787 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:15:04 crc kubenswrapper[4787]: I1203 17:15:04.765644 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:04 crc kubenswrapper[4787]: I1203 17:15:04.765739 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:04 crc kubenswrapper[4787]: I1203 17:15:04.765680 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:04 crc kubenswrapper[4787]: E1203 17:15:04.765912 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:15:04 crc kubenswrapper[4787]: E1203 17:15:04.766123 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:15:04 crc kubenswrapper[4787]: E1203 17:15:04.766277 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:15:05 crc kubenswrapper[4787]: I1203 17:15:05.765602 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:05 crc kubenswrapper[4787]: E1203 17:15:05.765828 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:15:06 crc kubenswrapper[4787]: I1203 17:15:06.765313 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:06 crc kubenswrapper[4787]: I1203 17:15:06.765366 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:06 crc kubenswrapper[4787]: E1203 17:15:06.765807 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:15:06 crc kubenswrapper[4787]: I1203 17:15:06.765410 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:06 crc kubenswrapper[4787]: E1203 17:15:06.765903 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:15:06 crc kubenswrapper[4787]: E1203 17:15:06.766278 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:15:07 crc kubenswrapper[4787]: I1203 17:15:07.765672 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:07 crc kubenswrapper[4787]: E1203 17:15:07.765806 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:15:08 crc kubenswrapper[4787]: I1203 17:15:08.765856 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:08 crc kubenswrapper[4787]: I1203 17:15:08.765939 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:08 crc kubenswrapper[4787]: I1203 17:15:08.765980 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:08 crc kubenswrapper[4787]: E1203 17:15:08.766069 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:15:08 crc kubenswrapper[4787]: E1203 17:15:08.766136 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:15:08 crc kubenswrapper[4787]: E1203 17:15:08.766298 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:15:08 crc kubenswrapper[4787]: E1203 17:15:08.848947 4787 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:15:09 crc kubenswrapper[4787]: I1203 17:15:09.764984 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:09 crc kubenswrapper[4787]: E1203 17:15:09.765285 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:15:09 crc kubenswrapper[4787]: I1203 17:15:09.765713 4787 scope.go:117] "RemoveContainer" containerID="59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8" Dec 03 17:15:10 crc kubenswrapper[4787]: I1203 17:15:10.471911 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xcz4f_bd5617c0-aa4f-4f21-a131-15af831725ec/kube-multus/1.log" Dec 03 17:15:10 crc kubenswrapper[4787]: I1203 17:15:10.472325 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xcz4f" event={"ID":"bd5617c0-aa4f-4f21-a131-15af831725ec","Type":"ContainerStarted","Data":"d24796ebc7e29b1d6ddb686edefd44466c78b350cb28a0f36ea374203323649a"} Dec 03 17:15:10 crc kubenswrapper[4787]: I1203 17:15:10.765876 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:10 crc kubenswrapper[4787]: I1203 17:15:10.765879 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:10 crc kubenswrapper[4787]: E1203 17:15:10.766106 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:15:10 crc kubenswrapper[4787]: E1203 17:15:10.766185 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:15:10 crc kubenswrapper[4787]: I1203 17:15:10.765903 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:10 crc kubenswrapper[4787]: E1203 17:15:10.766308 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:15:11 crc kubenswrapper[4787]: I1203 17:15:11.765328 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:11 crc kubenswrapper[4787]: E1203 17:15:11.765571 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:15:12 crc kubenswrapper[4787]: I1203 17:15:12.765605 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:12 crc kubenswrapper[4787]: I1203 17:15:12.765652 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:12 crc kubenswrapper[4787]: I1203 17:15:12.765730 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:12 crc kubenswrapper[4787]: E1203 17:15:12.765841 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:15:12 crc kubenswrapper[4787]: E1203 17:15:12.765910 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:15:12 crc kubenswrapper[4787]: E1203 17:15:12.766112 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:15:12 crc kubenswrapper[4787]: I1203 17:15:12.767212 4787 scope.go:117] "RemoveContainer" containerID="d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6" Dec 03 17:15:13 crc kubenswrapper[4787]: I1203 17:15:13.485830 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/3.log" Dec 03 17:15:13 crc kubenswrapper[4787]: I1203 17:15:13.488913 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerStarted","Data":"08b8f23afaf8a16b6613bf1e8f2bfea737d61a693f5a6f6c3c2213cf3202a093"} Dec 03 17:15:13 crc kubenswrapper[4787]: I1203 17:15:13.489267 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:15:13 crc kubenswrapper[4787]: I1203 17:15:13.629191 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podStartSLOduration=111.629159363 podStartE2EDuration="1m51.629159363s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:13.521210092 +0000 UTC m=+130.338681091" watchObservedRunningTime="2025-12-03 17:15:13.629159363 +0000 UTC m=+130.446630352" Dec 03 17:15:13 crc kubenswrapper[4787]: I1203 17:15:13.630720 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-m9tr6"] Dec 03 17:15:13 crc kubenswrapper[4787]: I1203 17:15:13.630864 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:13 crc kubenswrapper[4787]: E1203 17:15:13.631084 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:15:13 crc kubenswrapper[4787]: E1203 17:15:13.850478 4787 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:15:14 crc kubenswrapper[4787]: I1203 17:15:14.765766 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:14 crc kubenswrapper[4787]: E1203 17:15:14.766208 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:15:14 crc kubenswrapper[4787]: I1203 17:15:14.765844 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:14 crc kubenswrapper[4787]: I1203 17:15:14.765787 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:14 crc kubenswrapper[4787]: E1203 17:15:14.766287 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:15:14 crc kubenswrapper[4787]: E1203 17:15:14.766544 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:15:15 crc kubenswrapper[4787]: I1203 17:15:15.765600 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:15 crc kubenswrapper[4787]: E1203 17:15:15.765809 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:15:16 crc kubenswrapper[4787]: I1203 17:15:16.765240 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:16 crc kubenswrapper[4787]: I1203 17:15:16.765332 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:16 crc kubenswrapper[4787]: I1203 17:15:16.765297 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:16 crc kubenswrapper[4787]: E1203 17:15:16.765537 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:15:16 crc kubenswrapper[4787]: E1203 17:15:16.765752 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:15:16 crc kubenswrapper[4787]: E1203 17:15:16.765873 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:15:17 crc kubenswrapper[4787]: I1203 17:15:17.765532 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:17 crc kubenswrapper[4787]: E1203 17:15:17.765719 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-m9tr6" podUID="50c2569b-2a14-4112-82e4-afc683aa36a7" Dec 03 17:15:18 crc kubenswrapper[4787]: I1203 17:15:18.765697 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:18 crc kubenswrapper[4787]: I1203 17:15:18.765712 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:18 crc kubenswrapper[4787]: E1203 17:15:18.765916 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 03 17:15:18 crc kubenswrapper[4787]: E1203 17:15:18.766095 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 03 17:15:18 crc kubenswrapper[4787]: I1203 17:15:18.765727 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:18 crc kubenswrapper[4787]: E1203 17:15:18.766239 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 03 17:15:19 crc kubenswrapper[4787]: I1203 17:15:19.765101 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:19 crc kubenswrapper[4787]: I1203 17:15:19.767221 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 17:15:19 crc kubenswrapper[4787]: I1203 17:15:19.767685 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 17:15:20 crc kubenswrapper[4787]: I1203 17:15:20.765897 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:20 crc kubenswrapper[4787]: I1203 17:15:20.765961 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:20 crc kubenswrapper[4787]: I1203 17:15:20.765897 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:20 crc kubenswrapper[4787]: I1203 17:15:20.767689 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 17:15:20 crc kubenswrapper[4787]: I1203 17:15:20.768447 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 17:15:20 crc kubenswrapper[4787]: I1203 17:15:20.768922 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 17:15:20 crc kubenswrapper[4787]: I1203 17:15:20.769153 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.410594 4787 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.471687 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lwzb2"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.473050 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.473120 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.474174 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.474883 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.475808 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.477015 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-t88fj"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.477777 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.481528 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sm94s"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.483267 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.492662 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.493437 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.493820 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.494303 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.494643 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.494974 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.495396 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.495682 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496071 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496315 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496479 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496627 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496642 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496742 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496748 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496872 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496936 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496954 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496885 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.496315 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.497133 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.497270 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.500500 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bhg5b"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.502885 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rhg28"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.503286 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.503800 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.504246 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-xft52"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.504644 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xft52" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.505067 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pjpcp"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.505160 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.505344 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.505449 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.505781 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.516859 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.517278 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.517437 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.517715 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.517978 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.518404 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.518905 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.519103 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.519101 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-nm7m8"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.519268 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.519438 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.519645 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.519893 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.533457 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.535161 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.535967 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.536120 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.536524 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.536564 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.536687 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.536857 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.537141 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.537221 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.537251 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.537629 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.537926 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.538383 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.538400 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.539328 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.540392 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-f2vjk"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.541625 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.559859 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.560233 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.564849 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.565297 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.565771 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.567496 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.570332 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gdftf"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.571676 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.571754 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.574732 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.576317 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.582310 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.586096 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.586368 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.586526 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.586742 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.587545 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.588223 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.588437 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.588669 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.589197 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.589364 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.589490 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.589584 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.589849 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.589986 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.590187 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.590212 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.590461 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.590585 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.590731 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.591790 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.591895 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.592033 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.592149 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.592804 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.592870 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-x2wsr"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.592976 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593112 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593179 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593238 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593362 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593404 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593469 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593482 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593568 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593600 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593628 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593786 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593797 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594178 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnp85\" (UniqueName: \"kubernetes.io/projected/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-kube-api-access-lnp85\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594287 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-config\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593820 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594396 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-trusted-ca\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594587 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-image-import-ca\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594680 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea7b4185-4d6b-4d35-af22-252dc98d140d-config\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594755 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025afbb8-3e97-4e1e-a1c8-c50af81daba2-serving-cert\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594094 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594609 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594875 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595321 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595373 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-policies\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595394 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595417 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-etcd-serving-ca\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595437 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-etcd-client\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594719 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595457 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p66pk\" (UniqueName: \"kubernetes.io/projected/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-kube-api-access-p66pk\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595479 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-serving-cert\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595499 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595522 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22nkc\" (UniqueName: \"kubernetes.io/projected/025afbb8-3e97-4e1e-a1c8-c50af81daba2-kube-api-access-22nkc\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594285 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595542 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595565 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c2293a6-207e-4334-bb36-6e717de13ea0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-p875r\" (UID: \"8c2293a6-207e-4334-bb36-6e717de13ea0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594729 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595614 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-client-ca\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595637 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595657 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-serving-cert\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.593870 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595677 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-serving-cert\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595696 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-service-ca\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595717 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ea7b4185-4d6b-4d35-af22-252dc98d140d-machine-approver-tls\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595737 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594740 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595811 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.594933 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595895 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlhj9\" (UniqueName: \"kubernetes.io/projected/ea7b4185-4d6b-4d35-af22-252dc98d140d-kube-api-access-jlhj9\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595921 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9d570d04-9282-45fd-848b-7752fa9062a0-etcd-client\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595945 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595972 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8df314ae-77d5-48cb-9d05-6737236e0a9f-serving-cert\") pod \"openshift-config-operator-7777fb866f-fpjsk\" (UID: \"8df314ae-77d5-48cb-9d05-6737236e0a9f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.595994 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-serving-cert\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596260 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596546 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025afbb8-3e97-4e1e-a1c8-c50af81daba2-config\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596579 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596604 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9d570d04-9282-45fd-848b-7752fa9062a0-node-pullsecrets\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596624 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ml4h\" (UniqueName: \"kubernetes.io/projected/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-kube-api-access-4ml4h\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596667 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-encryption-config\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596751 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xrzg\" (UniqueName: \"kubernetes.io/projected/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-kube-api-access-8xrzg\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596771 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-trusted-ca-bundle\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596836 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/025afbb8-3e97-4e1e-a1c8-c50af81daba2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596859 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-oauth-serving-cert\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596899 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c2293a6-207e-4334-bb36-6e717de13ea0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-p875r\" (UID: \"8c2293a6-207e-4334-bb36-6e717de13ea0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.596933 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8df314ae-77d5-48cb-9d05-6737236e0a9f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fpjsk\" (UID: \"8df314ae-77d5-48cb-9d05-6737236e0a9f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597045 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597075 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-console-config\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597140 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597165 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-config\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597187 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-dir\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597207 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-oauth-config\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597255 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzgfv\" (UniqueName: \"kubernetes.io/projected/a66535aa-a358-4570-ae9f-5dac58121ef2-kube-api-access-bzgfv\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597277 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgkxg\" (UniqueName: \"kubernetes.io/projected/ade4b5ab-2584-49e9-9ece-a02a16d79a99-kube-api-access-rgkxg\") pod \"downloads-7954f5f757-xft52\" (UID: \"ade4b5ab-2584-49e9-9ece-a02a16d79a99\") " pod="openshift-console/downloads-7954f5f757-xft52" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597307 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48860c61-124e-43cf-9cf6-fd36f33866f5-config\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597330 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5s7wj\" (UniqueName: \"kubernetes.io/projected/4cf3d039-080b-4a38-a7e5-0a1e02f9f526-kube-api-access-5s7wj\") pod \"cluster-samples-operator-665b6dd947-rcs6j\" (UID: \"4cf3d039-080b-4a38-a7e5-0a1e02f9f526\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597384 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxlxr\" (UniqueName: \"kubernetes.io/projected/9d570d04-9282-45fd-848b-7752fa9062a0-kube-api-access-gxlxr\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597404 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597434 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kjgws\" (UniqueName: \"kubernetes.io/projected/8c2293a6-207e-4334-bb36-6e717de13ea0-kube-api-access-kjgws\") pod \"openshift-controller-manager-operator-756b6f6bc6-p875r\" (UID: \"8c2293a6-207e-4334-bb36-6e717de13ea0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597470 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jlqrb\" (UniqueName: \"kubernetes.io/projected/670e0167-14c0-4547-9539-f303e18c33f3-kube-api-access-jlqrb\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597505 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-config\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597539 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/025afbb8-3e97-4e1e-a1c8-c50af81daba2-service-ca-bundle\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597557 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-audit\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597586 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9d570d04-9282-45fd-848b-7752fa9062a0-encryption-config\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597604 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597626 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-client-ca\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597656 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-audit-policies\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597675 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2hj7\" (UniqueName: \"kubernetes.io/projected/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-kube-api-access-f2hj7\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597696 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d570d04-9282-45fd-848b-7752fa9062a0-serving-cert\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597726 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597745 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/48860c61-124e-43cf-9cf6-fd36f33866f5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597761 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-audit-dir\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597798 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5lmf\" (UniqueName: \"kubernetes.io/projected/8df314ae-77d5-48cb-9d05-6737236e0a9f-kube-api-access-x5lmf\") pod \"openshift-config-operator-7777fb866f-fpjsk\" (UID: \"8df314ae-77d5-48cb-9d05-6737236e0a9f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597824 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597841 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4cf3d039-080b-4a38-a7e5-0a1e02f9f526-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rcs6j\" (UID: \"4cf3d039-080b-4a38-a7e5-0a1e02f9f526\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597856 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/670e0167-14c0-4547-9539-f303e18c33f3-serving-cert\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597930 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597953 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9d570d04-9282-45fd-848b-7752fa9062a0-audit-dir\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597972 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-config\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.597992 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48860c61-124e-43cf-9cf6-fd36f33866f5-images\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.598150 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vbhm\" (UniqueName: \"kubernetes.io/projected/48860c61-124e-43cf-9cf6-fd36f33866f5-kube-api-access-6vbhm\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.598182 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ea7b4185-4d6b-4d35-af22-252dc98d140d-auth-proxy-config\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.598552 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.600191 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.601794 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.609979 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.628136 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.628841 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.628949 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.630062 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.650343 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.650989 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.652689 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.652820 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.656949 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.658670 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.661686 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dw4qj"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.661917 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.662501 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.663550 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lwzb2"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.666193 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.667402 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-t88fj"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.667557 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.676483 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.686069 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.687047 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.687219 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.687933 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.688168 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.688546 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.696430 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.700425 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-dir\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.701394 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-oauth-config\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.701506 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzgfv\" (UniqueName: \"kubernetes.io/projected/a66535aa-a358-4570-ae9f-5dac58121ef2-kube-api-access-bzgfv\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.701603 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgkxg\" (UniqueName: \"kubernetes.io/projected/ade4b5ab-2584-49e9-9ece-a02a16d79a99-kube-api-access-rgkxg\") pod \"downloads-7954f5f757-xft52\" (UID: \"ade4b5ab-2584-49e9-9ece-a02a16d79a99\") " pod="openshift-console/downloads-7954f5f757-xft52" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.701755 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.701910 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48860c61-124e-43cf-9cf6-fd36f33866f5-config\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.702066 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5s7wj\" (UniqueName: \"kubernetes.io/projected/4cf3d039-080b-4a38-a7e5-0a1e02f9f526-kube-api-access-5s7wj\") pod \"cluster-samples-operator-665b6dd947-rcs6j\" (UID: \"4cf3d039-080b-4a38-a7e5-0a1e02f9f526\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.702202 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxlxr\" (UniqueName: \"kubernetes.io/projected/9d570d04-9282-45fd-848b-7752fa9062a0-kube-api-access-gxlxr\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.702290 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kjgws\" (UniqueName: \"kubernetes.io/projected/8c2293a6-207e-4334-bb36-6e717de13ea0-kube-api-access-kjgws\") pod \"openshift-controller-manager-operator-756b6f6bc6-p875r\" (UID: \"8c2293a6-207e-4334-bb36-6e717de13ea0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.702368 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlqrb\" (UniqueName: \"kubernetes.io/projected/670e0167-14c0-4547-9539-f303e18c33f3-kube-api-access-jlqrb\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.702487 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/12108aae-ded8-4fc7-b763-bd6bfd13aa81-etcd-client\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.701012 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-dir\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.702624 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/12108aae-ded8-4fc7-b763-bd6bfd13aa81-etcd-service-ca\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.702717 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.706651 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-config\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.707078 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/025afbb8-3e97-4e1e-a1c8-c50af81daba2-service-ca-bundle\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.705855 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.706549 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.703534 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-46g9l"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.704760 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/48860c61-124e-43cf-9cf6-fd36f33866f5-config\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.708056 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-config\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.708145 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/025afbb8-3e97-4e1e-a1c8-c50af81daba2-service-ca-bundle\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.708205 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xft52"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.708239 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.708334 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.708706 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709105 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-audit\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709157 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9d570d04-9282-45fd-848b-7752fa9062a0-encryption-config\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709186 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-client-ca\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709211 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-audit-policies\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709242 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2hj7\" (UniqueName: \"kubernetes.io/projected/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-kube-api-access-f2hj7\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709274 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f637fc79-77c3-492a-8d39-c5c8454fd322-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mplbm\" (UID: \"f637fc79-77c3-492a-8d39-c5c8454fd322\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709302 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d570d04-9282-45fd-848b-7752fa9062a0-serving-cert\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709325 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709350 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/48860c61-124e-43cf-9cf6-fd36f33866f5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709373 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f637fc79-77c3-492a-8d39-c5c8454fd322-config\") pod \"kube-apiserver-operator-766d6c64bb-mplbm\" (UID: \"f637fc79-77c3-492a-8d39-c5c8454fd322\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709397 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-audit-dir\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709420 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5lmf\" (UniqueName: \"kubernetes.io/projected/8df314ae-77d5-48cb-9d05-6737236e0a9f-kube-api-access-x5lmf\") pod \"openshift-config-operator-7777fb866f-fpjsk\" (UID: \"8df314ae-77d5-48cb-9d05-6737236e0a9f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709451 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709477 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4cf3d039-080b-4a38-a7e5-0a1e02f9f526-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rcs6j\" (UID: \"4cf3d039-080b-4a38-a7e5-0a1e02f9f526\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709499 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/670e0167-14c0-4547-9539-f303e18c33f3-serving-cert\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709526 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ktpbv\" (UniqueName: \"kubernetes.io/projected/12108aae-ded8-4fc7-b763-bd6bfd13aa81-kube-api-access-ktpbv\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709551 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709572 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9d570d04-9282-45fd-848b-7752fa9062a0-audit-dir\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709599 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrvgn\" (UniqueName: \"kubernetes.io/projected/27bf1dec-a834-41e8-acdd-e5047002ed0b-kube-api-access-mrvgn\") pod \"dns-operator-744455d44c-gdftf\" (UID: \"27bf1dec-a834-41e8-acdd-e5047002ed0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709621 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12108aae-ded8-4fc7-b763-bd6bfd13aa81-config\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709645 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-config\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709665 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48860c61-124e-43cf-9cf6-fd36f33866f5-images\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709685 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vbhm\" (UniqueName: \"kubernetes.io/projected/48860c61-124e-43cf-9cf6-fd36f33866f5-kube-api-access-6vbhm\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709709 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ea7b4185-4d6b-4d35-af22-252dc98d140d-auth-proxy-config\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709736 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709757 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnp85\" (UniqueName: \"kubernetes.io/projected/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-kube-api-access-lnp85\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709778 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/12108aae-ded8-4fc7-b763-bd6bfd13aa81-etcd-ca\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709816 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-config\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709836 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-trusted-ca\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709856 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709881 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-image-import-ca\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709904 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea7b4185-4d6b-4d35-af22-252dc98d140d-config\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709934 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025afbb8-3e97-4e1e-a1c8-c50af81daba2-serving-cert\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709958 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.709979 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/27bf1dec-a834-41e8-acdd-e5047002ed0b-metrics-tls\") pod \"dns-operator-744455d44c-gdftf\" (UID: \"27bf1dec-a834-41e8-acdd-e5047002ed0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710006 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p66pk\" (UniqueName: \"kubernetes.io/projected/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-kube-api-access-p66pk\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710064 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12108aae-ded8-4fc7-b763-bd6bfd13aa81-serving-cert\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710091 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-policies\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710117 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710141 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-etcd-serving-ca\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710162 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-etcd-client\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710185 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710214 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-serving-cert\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710238 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710264 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22nkc\" (UniqueName: \"kubernetes.io/projected/025afbb8-3e97-4e1e-a1c8-c50af81daba2-kube-api-access-22nkc\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710291 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-service-ca\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710314 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c2293a6-207e-4334-bb36-6e717de13ea0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-p875r\" (UID: \"8c2293a6-207e-4334-bb36-6e717de13ea0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710341 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-client-ca\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710366 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710390 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-serving-cert\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710411 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-serving-cert\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710436 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ea7b4185-4d6b-4d35-af22-252dc98d140d-machine-approver-tls\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710459 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710493 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f637fc79-77c3-492a-8d39-c5c8454fd322-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mplbm\" (UID: \"f637fc79-77c3-492a-8d39-c5c8454fd322\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710522 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710550 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jlhj9\" (UniqueName: \"kubernetes.io/projected/ea7b4185-4d6b-4d35-af22-252dc98d140d-kube-api-access-jlhj9\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710577 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9d570d04-9282-45fd-848b-7752fa9062a0-etcd-client\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710602 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710624 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8df314ae-77d5-48cb-9d05-6737236e0a9f-serving-cert\") pod \"openshift-config-operator-7777fb866f-fpjsk\" (UID: \"8df314ae-77d5-48cb-9d05-6737236e0a9f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710648 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ml4h\" (UniqueName: \"kubernetes.io/projected/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-kube-api-access-4ml4h\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710671 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-serving-cert\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710698 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025afbb8-3e97-4e1e-a1c8-c50af81daba2-config\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710724 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710749 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9d570d04-9282-45fd-848b-7752fa9062a0-node-pullsecrets\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710793 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-encryption-config\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710816 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-trusted-ca-bundle\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710850 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xrzg\" (UniqueName: \"kubernetes.io/projected/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-kube-api-access-8xrzg\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710872 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8df314ae-77d5-48cb-9d05-6737236e0a9f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fpjsk\" (UID: \"8df314ae-77d5-48cb-9d05-6737236e0a9f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710896 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/025afbb8-3e97-4e1e-a1c8-c50af81daba2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710917 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-oauth-serving-cert\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710937 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c2293a6-207e-4334-bb36-6e717de13ea0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-p875r\" (UID: \"8c2293a6-207e-4334-bb36-6e717de13ea0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.710985 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.711006 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-console-config\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.711067 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.711088 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-config\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.711749 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-config\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.711822 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/9d570d04-9282-45fd-848b-7752fa9062a0-audit-dir\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.711948 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-config\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.713186 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/48860c61-124e-43cf-9cf6-fd36f33866f5-images\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.714255 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.714258 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-oauth-config\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.714350 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/670e0167-14c0-4547-9539-f303e18c33f3-serving-cert\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.714539 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-audit-policies\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.714909 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/ea7b4185-4d6b-4d35-af22-252dc98d140d-auth-proxy-config\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.715148 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-audit-dir\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.715687 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-trusted-ca-bundle\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.715917 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-client-ca\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.715988 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.716736 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-config\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.716791 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/9d570d04-9282-45fd-848b-7752fa9062a0-node-pullsecrets\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.718181 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-serving-cert\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.718640 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-console-config\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.719430 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.719680 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.719679 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.720894 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.720998 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.721133 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-encryption-config\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.721441 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.721801 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-oauth-serving-cert\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.722137 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-trusted-ca-bundle\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.722428 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d570d04-9282-45fd-848b-7752fa9062a0-serving-cert\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.723008 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ea7b4185-4d6b-4d35-af22-252dc98d140d-config\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.723469 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.724119 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.724143 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.724225 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/025afbb8-3e97-4e1e-a1c8-c50af81daba2-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.724470 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8c2293a6-207e-4334-bb36-6e717de13ea0-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-p875r\" (UID: \"8c2293a6-207e-4334-bb36-6e717de13ea0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.724566 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/025afbb8-3e97-4e1e-a1c8-c50af81daba2-config\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.724945 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.725279 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8df314ae-77d5-48cb-9d05-6737236e0a9f-available-featuregates\") pod \"openshift-config-operator-7777fb866f-fpjsk\" (UID: \"8df314ae-77d5-48cb-9d05-6737236e0a9f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.725361 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-trusted-ca\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.725749 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.725928 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/48860c61-124e-43cf-9cf6-fd36f33866f5-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.726527 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-service-ca\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.726574 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.726559 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-audit\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.726690 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.726963 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/9d570d04-9282-45fd-848b-7752fa9062a0-etcd-client\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.727175 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-policies\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.727244 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/9d570d04-9282-45fd-848b-7752fa9062a0-encryption-config\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.727674 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-etcd-serving-ca\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.727910 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.728545 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-etcd-client\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.728592 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-client-ca\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.728748 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/9d570d04-9282-45fd-848b-7752fa9062a0-image-import-ca\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.728950 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8df314ae-77d5-48cb-9d05-6737236e0a9f-serving-cert\") pod \"openshift-config-operator-7777fb866f-fpjsk\" (UID: \"8df314ae-77d5-48cb-9d05-6737236e0a9f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.729065 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/4cf3d039-080b-4a38-a7e5-0a1e02f9f526-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rcs6j\" (UID: \"4cf3d039-080b-4a38-a7e5-0a1e02f9f526\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.729239 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/025afbb8-3e97-4e1e-a1c8-c50af81daba2-serving-cert\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.729328 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.730147 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.730616 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.731226 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/ea7b4185-4d6b-4d35-af22-252dc98d140d-machine-approver-tls\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.731542 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.731863 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rb22r"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.732418 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.733098 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.733869 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.735406 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bl7vk"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.735598 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-serving-cert\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.735797 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.737751 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.737973 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-serving-cert\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.738183 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.738408 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.739599 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.739715 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-serving-cert\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.740580 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8c2293a6-207e-4334-bb36-6e717de13ea0-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-p875r\" (UID: \"8c2293a6-207e-4334-bb36-6e717de13ea0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.747946 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.748372 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.750730 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.750931 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.751679 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.754065 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.755336 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.756578 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.758320 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-x7g9b"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.759082 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.760462 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.761886 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.767072 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-nm7m8"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.768888 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gdftf"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.770402 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-7m57w"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.771078 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7m57w" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.771903 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.772780 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bhg5b"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.773882 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rhg28"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.775455 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.779293 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.780214 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.780931 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sm94s"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.782979 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.784255 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.785952 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dw4qj"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.787580 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.789235 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pjpcp"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.790841 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.792416 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.794102 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.795127 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.795653 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-f2vjk"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.797291 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.798949 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-jjvtn"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.799908 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-jjvtn" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.800509 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-vjhsg"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.801332 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.802032 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.802947 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.803976 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.804993 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.805984 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bl7vk"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.806975 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.808247 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-46g9l"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.809391 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.810340 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rb22r"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.811369 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-x7g9b"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812069 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/27bf1dec-a834-41e8-acdd-e5047002ed0b-metrics-tls\") pod \"dns-operator-744455d44c-gdftf\" (UID: \"27bf1dec-a834-41e8-acdd-e5047002ed0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812154 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12108aae-ded8-4fc7-b763-bd6bfd13aa81-serving-cert\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812198 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f637fc79-77c3-492a-8d39-c5c8454fd322-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mplbm\" (UID: \"f637fc79-77c3-492a-8d39-c5c8454fd322\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812353 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/12108aae-ded8-4fc7-b763-bd6bfd13aa81-etcd-client\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812385 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/12108aae-ded8-4fc7-b763-bd6bfd13aa81-etcd-service-ca\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812411 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f637fc79-77c3-492a-8d39-c5c8454fd322-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mplbm\" (UID: \"f637fc79-77c3-492a-8d39-c5c8454fd322\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812429 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f637fc79-77c3-492a-8d39-c5c8454fd322-config\") pod \"kube-apiserver-operator-766d6c64bb-mplbm\" (UID: \"f637fc79-77c3-492a-8d39-c5c8454fd322\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812459 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrvgn\" (UniqueName: \"kubernetes.io/projected/27bf1dec-a834-41e8-acdd-e5047002ed0b-kube-api-access-mrvgn\") pod \"dns-operator-744455d44c-gdftf\" (UID: \"27bf1dec-a834-41e8-acdd-e5047002ed0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812479 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12108aae-ded8-4fc7-b763-bd6bfd13aa81-config\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812496 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ktpbv\" (UniqueName: \"kubernetes.io/projected/12108aae-ded8-4fc7-b763-bd6bfd13aa81-kube-api-access-ktpbv\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812531 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/12108aae-ded8-4fc7-b763-bd6bfd13aa81-etcd-ca\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.812709 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.813414 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.813449 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/12108aae-ded8-4fc7-b763-bd6bfd13aa81-etcd-service-ca\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.813697 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/12108aae-ded8-4fc7-b763-bd6bfd13aa81-config\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.813959 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/12108aae-ded8-4fc7-b763-bd6bfd13aa81-etcd-ca\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.815075 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.815329 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7m57w"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.818495 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/12108aae-ded8-4fc7-b763-bd6bfd13aa81-serving-cert\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.821856 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vjhsg"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.821910 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cggxn"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.823189 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/27bf1dec-a834-41e8-acdd-e5047002ed0b-metrics-tls\") pod \"dns-operator-744455d44c-gdftf\" (UID: \"27bf1dec-a834-41e8-acdd-e5047002ed0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.825486 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cggxn"] Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.825638 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.828276 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/12108aae-ded8-4fc7-b763-bd6bfd13aa81-etcd-client\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.855279 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.875119 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.895566 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.915982 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.936125 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.955868 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.975758 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 17:15:28 crc kubenswrapper[4787]: I1203 17:15:28.995507 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.015827 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.025169 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f637fc79-77c3-492a-8d39-c5c8454fd322-config\") pod \"kube-apiserver-operator-766d6c64bb-mplbm\" (UID: \"f637fc79-77c3-492a-8d39-c5c8454fd322\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.036166 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.046132 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f637fc79-77c3-492a-8d39-c5c8454fd322-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-mplbm\" (UID: \"f637fc79-77c3-492a-8d39-c5c8454fd322\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.055874 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.076242 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.096176 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.116047 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.156423 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.178533 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.196363 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.217078 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.236694 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.255252 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.277242 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.295881 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.315921 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.336969 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.356399 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.377746 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.413844 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgkxg\" (UniqueName: \"kubernetes.io/projected/ade4b5ab-2584-49e9-9ece-a02a16d79a99-kube-api-access-rgkxg\") pod \"downloads-7954f5f757-xft52\" (UID: \"ade4b5ab-2584-49e9-9ece-a02a16d79a99\") " pod="openshift-console/downloads-7954f5f757-xft52" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.431279 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzgfv\" (UniqueName: \"kubernetes.io/projected/a66535aa-a358-4570-ae9f-5dac58121ef2-kube-api-access-bzgfv\") pod \"console-f9d7485db-nm7m8\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.455646 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxlxr\" (UniqueName: \"kubernetes.io/projected/9d570d04-9282-45fd-848b-7752fa9062a0-kube-api-access-gxlxr\") pod \"apiserver-76f77b778f-lwzb2\" (UID: \"9d570d04-9282-45fd-848b-7752fa9062a0\") " pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.474196 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5s7wj\" (UniqueName: \"kubernetes.io/projected/4cf3d039-080b-4a38-a7e5-0a1e02f9f526-kube-api-access-5s7wj\") pod \"cluster-samples-operator-665b6dd947-rcs6j\" (UID: \"4cf3d039-080b-4a38-a7e5-0a1e02f9f526\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.493173 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlqrb\" (UniqueName: \"kubernetes.io/projected/670e0167-14c0-4547-9539-f303e18c33f3-kube-api-access-jlqrb\") pod \"route-controller-manager-6576b87f9c-m4lbg\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.514552 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kjgws\" (UniqueName: \"kubernetes.io/projected/8c2293a6-207e-4334-bb36-6e717de13ea0-kube-api-access-kjgws\") pod \"openshift-controller-manager-operator-756b6f6bc6-p875r\" (UID: \"8c2293a6-207e-4334-bb36-6e717de13ea0\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.517181 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.536608 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.550821 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xft52" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.557321 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.575747 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.587622 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.595664 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.610812 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.616220 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.636362 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.675654 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.685339 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnp85\" (UniqueName: \"kubernetes.io/projected/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-kube-api-access-lnp85\") pod \"controller-manager-879f6c89f-t88fj\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.693659 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vbhm\" (UniqueName: \"kubernetes.io/projected/48860c61-124e-43cf-9cf6-fd36f33866f5-kube-api-access-6vbhm\") pod \"machine-api-operator-5694c8668f-sm94s\" (UID: \"48860c61-124e-43cf-9cf6-fd36f33866f5\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.718622 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2hj7\" (UniqueName: \"kubernetes.io/projected/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-kube-api-access-f2hj7\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.723924 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.733541 4787 request.go:700] Waited for 1.015456023s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/serviceaccounts/oauth-openshift/token Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.736511 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5lmf\" (UniqueName: \"kubernetes.io/projected/8df314ae-77d5-48cb-9d05-6737236e0a9f-kube-api-access-x5lmf\") pod \"openshift-config-operator-7777fb866f-fpjsk\" (UID: \"8df314ae-77d5-48cb-9d05-6737236e0a9f\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.756854 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xrzg\" (UniqueName: \"kubernetes.io/projected/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-kube-api-access-8xrzg\") pod \"oauth-openshift-558db77b4-bhg5b\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.770850 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ml4h\" (UniqueName: \"kubernetes.io/projected/1c90cc96-7e0f-445a-9519-f1632e4b0b5d-kube-api-access-4ml4h\") pod \"apiserver-7bbb656c7d-wkg8k\" (UID: \"1c90cc96-7e0f-445a-9519-f1632e4b0b5d\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.778320 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.794623 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p66pk\" (UniqueName: \"kubernetes.io/projected/844a7007-5cb5-42b1-ada8-9b4faf3dddc1-kube-api-access-p66pk\") pod \"console-operator-58897d9998-pjpcp\" (UID: \"844a7007-5cb5-42b1-ada8-9b4faf3dddc1\") " pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.796878 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.810420 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jlhj9\" (UniqueName: \"kubernetes.io/projected/ea7b4185-4d6b-4d35-af22-252dc98d140d-kube-api-access-jlhj9\") pod \"machine-approver-56656f9798-86q7g\" (UID: \"ea7b4185-4d6b-4d35-af22-252dc98d140d\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.829671 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.831456 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.840399 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22nkc\" (UniqueName: \"kubernetes.io/projected/025afbb8-3e97-4e1e-a1c8-c50af81daba2-kube-api-access-22nkc\") pod \"authentication-operator-69f744f599-rhg28\" (UID: \"025afbb8-3e97-4e1e-a1c8-c50af81daba2\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.859072 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.860392 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.861930 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-l4sbk\" (UID: \"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.883637 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.883653 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.899459 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.903253 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.916594 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.935928 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.953405 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-nm7m8"] Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.956180 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 17:15:29 crc kubenswrapper[4787]: W1203 17:15:29.968063 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda66535aa_a358_4570_ae9f_5dac58121ef2.slice/crio-ec9e06a0eb112a7de167d3880784823d62b3a878cc9564bc48ece361534c05f6 WatchSource:0}: Error finding container ec9e06a0eb112a7de167d3880784823d62b3a878cc9564bc48ece361534c05f6: Status 404 returned error can't find the container with id ec9e06a0eb112a7de167d3880784823d62b3a878cc9564bc48ece361534c05f6 Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.968457 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.976501 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.985641 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.989961 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xft52"] Dec 03 17:15:29 crc kubenswrapper[4787]: I1203 17:15:29.999100 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.006912 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-lwzb2"] Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.016279 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.038081 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.056540 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.060253 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-t88fj"] Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.072323 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j"] Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.073337 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r"] Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.088257 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.097329 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.098912 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.117746 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.135856 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.155125 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.175486 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.180521 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-sm94s"] Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.199568 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 17:15:30 crc kubenswrapper[4787]: W1203 17:15:30.209946 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48860c61_124e_43cf_9cf6_fd36f33866f5.slice/crio-35a3e374a669c640202818f43f25e608793812f22bd0bf7b24f47ac1688793af WatchSource:0}: Error finding container 35a3e374a669c640202818f43f25e608793812f22bd0bf7b24f47ac1688793af: Status 404 returned error can't find the container with id 35a3e374a669c640202818f43f25e608793812f22bd0bf7b24f47ac1688793af Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.215511 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.235729 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.255824 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.282308 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.299179 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.319117 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.335940 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.357690 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.367440 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk"] Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.377013 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 17:15:30 crc kubenswrapper[4787]: W1203 17:15:30.379220 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cc4a1a3_bd7c_4ed7_a818_a1cef15dce4e.slice/crio-bd17d6e2e9047102fb66eda8eed88fe71fb16dc10f95389d8770550c45477dc1 WatchSource:0}: Error finding container bd17d6e2e9047102fb66eda8eed88fe71fb16dc10f95389d8770550c45477dc1: Status 404 returned error can't find the container with id bd17d6e2e9047102fb66eda8eed88fe71fb16dc10f95389d8770550c45477dc1 Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.394631 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-pjpcp"] Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.395471 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 17:15:30 crc kubenswrapper[4787]: W1203 17:15:30.406164 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod844a7007_5cb5_42b1_ada8_9b4faf3dddc1.slice/crio-0afc3963806cd638babb291b7d8f6dc765e3527b442c3ae401fc6eb44ee28b8c WatchSource:0}: Error finding container 0afc3963806cd638babb291b7d8f6dc765e3527b442c3ae401fc6eb44ee28b8c: Status 404 returned error can't find the container with id 0afc3963806cd638babb291b7d8f6dc765e3527b442c3ae401fc6eb44ee28b8c Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.415641 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.435559 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.455341 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.477816 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk"] Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.488414 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.496211 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.515073 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-rhg28"] Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.515696 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 17:15:30 crc kubenswrapper[4787]: W1203 17:15:30.530492 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod025afbb8_3e97_4e1e_a1c8_c50af81daba2.slice/crio-436b83035338535b2a406a94956bb2fda1dab06817f8cbf3b21acd18056cb133 WatchSource:0}: Error finding container 436b83035338535b2a406a94956bb2fda1dab06817f8cbf3b21acd18056cb133: Status 404 returned error can't find the container with id 436b83035338535b2a406a94956bb2fda1dab06817f8cbf3b21acd18056cb133 Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.530539 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k"] Dec 03 17:15:30 crc kubenswrapper[4787]: W1203 17:15:30.530822 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8df314ae_77d5_48cb_9d05_6737236e0a9f.slice/crio-822e10fbaa357f7a12b45e113ff030139fe7ae0b06fb67e776f864a942dfdf0f WatchSource:0}: Error finding container 822e10fbaa357f7a12b45e113ff030139fe7ae0b06fb67e776f864a942dfdf0f: Status 404 returned error can't find the container with id 822e10fbaa357f7a12b45e113ff030139fe7ae0b06fb67e776f864a942dfdf0f Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.538788 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.539209 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bhg5b"] Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.553590 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg"] Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.554842 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 17:15:30 crc kubenswrapper[4787]: W1203 17:15:30.572628 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode19b19ce_efa6_4d43_af6c_bb6aed7c232d.slice/crio-f778721fe4300af26b049636592d0f61b4792714f47688c6e65026e89ce6a9d9 WatchSource:0}: Error finding container f778721fe4300af26b049636592d0f61b4792714f47688c6e65026e89ce6a9d9: Status 404 returned error can't find the container with id f778721fe4300af26b049636592d0f61b4792714f47688c6e65026e89ce6a9d9 Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.576844 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.585623 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pjpcp" event={"ID":"844a7007-5cb5-42b1-ada8-9b4faf3dddc1","Type":"ContainerStarted","Data":"0afc3963806cd638babb291b7d8f6dc765e3527b442c3ae401fc6eb44ee28b8c"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.587385 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nm7m8" event={"ID":"a66535aa-a358-4570-ae9f-5dac58121ef2","Type":"ContainerStarted","Data":"7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.587437 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nm7m8" event={"ID":"a66535aa-a358-4570-ae9f-5dac58121ef2","Type":"ContainerStarted","Data":"ec9e06a0eb112a7de167d3880784823d62b3a878cc9564bc48ece361534c05f6"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.596697 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.596761 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xft52" event={"ID":"ade4b5ab-2584-49e9-9ece-a02a16d79a99","Type":"ContainerStarted","Data":"dab2de7e9e6911013b3ad72fa9e946e07984139f54ddc2d0329e53481dfe7015"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.596812 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xft52" event={"ID":"ade4b5ab-2584-49e9-9ece-a02a16d79a99","Type":"ContainerStarted","Data":"0f436f285da3a2de7e0b57a5753216f683033d019ebfa44797590666143d4091"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.597077 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-xft52" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.606261 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" event={"ID":"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e","Type":"ContainerStarted","Data":"bd17d6e2e9047102fb66eda8eed88fe71fb16dc10f95389d8770550c45477dc1"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.609327 4787 patch_prober.go:28] interesting pod/downloads-7954f5f757-xft52 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.609378 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xft52" podUID="ade4b5ab-2584-49e9-9ece-a02a16d79a99" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.609649 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" event={"ID":"8df314ae-77d5-48cb-9d05-6737236e0a9f","Type":"ContainerStarted","Data":"822e10fbaa357f7a12b45e113ff030139fe7ae0b06fb67e776f864a942dfdf0f"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.615424 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.620163 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" event={"ID":"1c90cc96-7e0f-445a-9519-f1632e4b0b5d","Type":"ContainerStarted","Data":"ab1c62c3f0a88968cf444f39b39aed40aa0d7c4571becc63ef6fc0bb34853ab9"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.621816 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" event={"ID":"48860c61-124e-43cf-9cf6-fd36f33866f5","Type":"ContainerStarted","Data":"83587c638a016686620b2740bac1509afa8a5ea06c318040b95979509acd30e4"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.621838 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" event={"ID":"48860c61-124e-43cf-9cf6-fd36f33866f5","Type":"ContainerStarted","Data":"35a3e374a669c640202818f43f25e608793812f22bd0bf7b24f47ac1688793af"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.623317 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" event={"ID":"8c2293a6-207e-4334-bb36-6e717de13ea0","Type":"ContainerStarted","Data":"0d99b21e8209d0b14c359cb4ff8cb87586ee8b4c46766ceeabcf4a4801c74e47"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.623340 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" event={"ID":"8c2293a6-207e-4334-bb36-6e717de13ea0","Type":"ContainerStarted","Data":"e49d894cc17848fc9f9f369e22751953fbc7ad3f3d1e71e3dd3fff9adaa7aa46"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.626246 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" event={"ID":"4cf3d039-080b-4a38-a7e5-0a1e02f9f526","Type":"ContainerStarted","Data":"1611f6d9c0776307b40512a04941601a2c3904772f1a979c977b83fdfce6560e"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.626272 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" event={"ID":"4cf3d039-080b-4a38-a7e5-0a1e02f9f526","Type":"ContainerStarted","Data":"d15f4b7b7750682c99477be15246a3369f932cebd061b741c92870f3f9c930e1"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.628054 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" event={"ID":"ea7b4185-4d6b-4d35-af22-252dc98d140d","Type":"ContainerStarted","Data":"219135115a38c3a79422083d63bcbda044961019650f5d4be1e86f2cfb1b53c0"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.628450 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" event={"ID":"ea7b4185-4d6b-4d35-af22-252dc98d140d","Type":"ContainerStarted","Data":"4b3ac484a8d48a8f5b68d29982bfbc4060ddc06a24f967ac75ad8fef794b8660"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.630716 4787 generic.go:334] "Generic (PLEG): container finished" podID="9d570d04-9282-45fd-848b-7752fa9062a0" containerID="8ab4f5d5bafc2280abd2a2d339ec94c3478999b431851e96375ef1d2e6c0463c" exitCode=0 Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.630791 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" event={"ID":"9d570d04-9282-45fd-848b-7752fa9062a0","Type":"ContainerDied","Data":"8ab4f5d5bafc2280abd2a2d339ec94c3478999b431851e96375ef1d2e6c0463c"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.630816 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" event={"ID":"9d570d04-9282-45fd-848b-7752fa9062a0","Type":"ContainerStarted","Data":"3b10e54c647a3de938e414976ec75704e2366202e625492deabf78e900fb73a1"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.634113 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" event={"ID":"025afbb8-3e97-4e1e-a1c8-c50af81daba2","Type":"ContainerStarted","Data":"436b83035338535b2a406a94956bb2fda1dab06817f8cbf3b21acd18056cb133"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.635118 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" event={"ID":"e19b19ce-efa6-4d43-af6c-bb6aed7c232d","Type":"ContainerStarted","Data":"f778721fe4300af26b049636592d0f61b4792714f47688c6e65026e89ce6a9d9"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.637866 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" event={"ID":"f41a2379-8c92-4d95-96cc-3d466bdcbc1d","Type":"ContainerStarted","Data":"e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.637889 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" event={"ID":"f41a2379-8c92-4d95-96cc-3d466bdcbc1d","Type":"ContainerStarted","Data":"1954648f6ee937697936af552aa349f74fa246b05c951b9479143bbdb161926b"} Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.638642 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.655189 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.655924 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.668844 4787 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-t88fj container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" start-of-body= Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.668899 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" podUID="f41a2379-8c92-4d95-96cc-3d466bdcbc1d" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.7:8443/healthz\": dial tcp 10.217.0.7:8443: connect: connection refused" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.676420 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.695561 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.715264 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.733919 4787 request.go:700] Waited for 1.932361079s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-default-metrics-tls&limit=500&resourceVersion=0 Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.735969 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.755072 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.755504 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:30 crc kubenswrapper[4787]: E1203 17:15:30.755605 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:17:32.755584933 +0000 UTC m=+269.573055892 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.757306 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.757633 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.761157 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.763182 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.791084 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f637fc79-77c3-492a-8d39-c5c8454fd322-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-mplbm\" (UID: \"f637fc79-77c3-492a-8d39-c5c8454fd322\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.810861 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrvgn\" (UniqueName: \"kubernetes.io/projected/27bf1dec-a834-41e8-acdd-e5047002ed0b-kube-api-access-mrvgn\") pod \"dns-operator-744455d44c-gdftf\" (UID: \"27bf1dec-a834-41e8-acdd-e5047002ed0b\") " pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.832704 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ktpbv\" (UniqueName: \"kubernetes.io/projected/12108aae-ded8-4fc7-b763-bd6bfd13aa81-kube-api-access-ktpbv\") pod \"etcd-operator-b45778765-f2vjk\" (UID: \"12108aae-ded8-4fc7-b763-bd6bfd13aa81\") " pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.840152 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.857256 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.858853 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.858912 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.864205 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.865252 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.877651 4787 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.893131 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.903871 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.947963 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.961811 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2df4d7cb-f9bb-4ece-9b95-36027a01f017-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.961880 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/70235f57-caf7-48d6-ab8e-85230e423cd0-stats-auth\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.961949 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70235f57-caf7-48d6-ab8e-85230e423cd0-metrics-certs\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962007 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-trusted-ca\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962064 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-tls\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962080 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jvwvq\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-kube-api-access-jvwvq\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962118 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2df4d7cb-f9bb-4ece-9b95-36027a01f017-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962161 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-certificates\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962186 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2hvw5\" (UniqueName: \"kubernetes.io/projected/70235f57-caf7-48d6-ab8e-85230e423cd0-kube-api-access-2hvw5\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962243 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2f958667-8c81-40a0-97c0-0afafae369d2-metrics-tls\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962319 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962562 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-bound-sa-token\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962599 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2f958667-8c81-40a0-97c0-0afafae369d2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962617 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxlt8\" (UniqueName: \"kubernetes.io/projected/2f958667-8c81-40a0-97c0-0afafae369d2-kube-api-access-xxlt8\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962648 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f958667-8c81-40a0-97c0-0afafae369d2-trusted-ca\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962669 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70235f57-caf7-48d6-ab8e-85230e423cd0-service-ca-bundle\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:30 crc kubenswrapper[4787]: I1203 17:15:30.962816 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/70235f57-caf7-48d6-ab8e-85230e423cd0-default-certificate\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:30 crc kubenswrapper[4787]: E1203 17:15:30.969845 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:31.469829882 +0000 UTC m=+148.287300841 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.007563 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.007688 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.008237 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.068428 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069487 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/70235f57-caf7-48d6-ab8e-85230e423cd0-stats-auth\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069529 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f58901-2824-4993-a0cd-c16e5509894a-config-volume\") pod \"dns-default-vjhsg\" (UID: \"b3f58901-2824-4993-a0cd-c16e5509894a\") " pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069565 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c09d8632-626c-485d-9e74-65e651546d04-certs\") pod \"machine-config-server-jjvtn\" (UID: \"c09d8632-626c-485d-9e74-65e651546d04\") " pod="openshift-machine-config-operator/machine-config-server-jjvtn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069583 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-secret-volume\") pod \"collect-profiles-29413035-tghvb\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069608 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88a31888-8dbe-45d6-9ce9-0d9c65e9f914-srv-cert\") pod \"olm-operator-6b444d44fb-zcnjw\" (UID: \"88a31888-8dbe-45d6-9ce9-0d9c65e9f914\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069657 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70235f57-caf7-48d6-ab8e-85230e423cd0-metrics-certs\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069675 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-csi-data-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069718 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-registration-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069759 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89648a64-0aeb-48b0-ad91-3aac3b4cd5e7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rwvrf\" (UID: \"89648a64-0aeb-48b0-ad91-3aac3b4cd5e7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069822 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-trusted-ca\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069843 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b3f58901-2824-4993-a0cd-c16e5509894a-metrics-tls\") pod \"dns-default-vjhsg\" (UID: \"b3f58901-2824-4993-a0cd-c16e5509894a\") " pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069891 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c0c6d238-9876-4f67-b557-94ac4f209332-images\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069926 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ftwj\" (UniqueName: \"kubernetes.io/projected/aba59ac8-ce2a-4321-a2a3-d790af00541c-kube-api-access-5ftwj\") pod \"service-ca-operator-777779d784-46g9l\" (UID: \"aba59ac8-ce2a-4321-a2a3-d790af00541c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069958 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88a31888-8dbe-45d6-9ce9-0d9c65e9f914-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zcnjw\" (UID: \"88a31888-8dbe-45d6-9ce9-0d9c65e9f914\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.069980 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klrdt\" (UniqueName: \"kubernetes.io/projected/c09d8632-626c-485d-9e74-65e651546d04-kube-api-access-klrdt\") pod \"machine-config-server-jjvtn\" (UID: \"c09d8632-626c-485d-9e74-65e651546d04\") " pod="openshift-machine-config-operator/machine-config-server-jjvtn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070000 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb5lp\" (UniqueName: \"kubernetes.io/projected/684133b3-732f-47c6-b15d-108c3e37397f-kube-api-access-mb5lp\") pod \"marketplace-operator-79b997595-bl7vk\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070488 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df870037-226f-4286-b128-bbf81e4871b8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zk9gm\" (UID: \"df870037-226f-4286-b128-bbf81e4871b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070512 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tlzdz\" (UniqueName: \"kubernetes.io/projected/88a31888-8dbe-45d6-9ce9-0d9c65e9f914-kube-api-access-tlzdz\") pod \"olm-operator-6b444d44fb-zcnjw\" (UID: \"88a31888-8dbe-45d6-9ce9-0d9c65e9f914\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070560 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89648a64-0aeb-48b0-ad91-3aac3b4cd5e7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rwvrf\" (UID: \"89648a64-0aeb-48b0-ad91-3aac3b4cd5e7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070579 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/688cf6b7-0d4d-452f-b878-93bd8ff75408-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jgrtv\" (UID: \"688cf6b7-0d4d-452f-b878-93bd8ff75408\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070634 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7-cert\") pod \"ingress-canary-7m57w\" (UID: \"f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7\") " pod="openshift-ingress-canary/ingress-canary-7m57w" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070652 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-plugins-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070672 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65vms\" (UniqueName: \"kubernetes.io/projected/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-kube-api-access-65vms\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070712 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bl7vk\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070730 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cxmp\" (UniqueName: \"kubernetes.io/projected/df870037-226f-4286-b128-bbf81e4871b8-kube-api-access-8cxmp\") pod \"machine-config-controller-84d6567774-zk9gm\" (UID: \"df870037-226f-4286-b128-bbf81e4871b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070747 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2f958667-8c81-40a0-97c0-0afafae369d2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070769 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxlt8\" (UniqueName: \"kubernetes.io/projected/2f958667-8c81-40a0-97c0-0afafae369d2-kube-api-access-xxlt8\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070812 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70235f57-caf7-48d6-ab8e-85230e423cd0-service-ca-bundle\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070844 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1af7cc12-ae37-4010-9d9c-28fb02ca03b4-profile-collector-cert\") pod \"catalog-operator-68c6474976-cj96h\" (UID: \"1af7cc12-ae37-4010-9d9c-28fb02ca03b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070900 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aba59ac8-ce2a-4321-a2a3-d790af00541c-serving-cert\") pod \"service-ca-operator-777779d784-46g9l\" (UID: \"aba59ac8-ce2a-4321-a2a3-d790af00541c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070922 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/552d32f0-c753-4e7d-b0e7-1707e8fa8d26-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gg7bd\" (UID: \"552d32f0-c753-4e7d-b0e7-1707e8fa8d26\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070967 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0466f2b9-1263-4369-80f0-364304f1f2f5-tmpfs\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.070985 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0466f2b9-1263-4369-80f0-364304f1f2f5-webhook-cert\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.071077 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1af7cc12-ae37-4010-9d9c-28fb02ca03b4-srv-cert\") pod \"catalog-operator-68c6474976-cj96h\" (UID: \"1af7cc12-ae37-4010-9d9c-28fb02ca03b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.071100 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/339b0916-61a9-4e97-9fd3-d3a8dc22d23f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-swngg\" (UID: \"339b0916-61a9-4e97-9fd3-d3a8dc22d23f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.071159 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4zlf\" (UniqueName: \"kubernetes.io/projected/89648a64-0aeb-48b0-ad91-3aac3b4cd5e7-kube-api-access-b4zlf\") pod \"openshift-apiserver-operator-796bbdcf4f-rwvrf\" (UID: \"89648a64-0aeb-48b0-ad91-3aac3b4cd5e7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.071187 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2df4d7cb-f9bb-4ece-9b95-36027a01f017-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.071209 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0466f2b9-1263-4369-80f0-364304f1f2f5-apiservice-cert\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.071263 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcnc5\" (UniqueName: \"kubernetes.io/projected/e87eccc7-4ea1-4cfe-a219-e8d727fd2051-kube-api-access-hcnc5\") pod \"service-ca-9c57cc56f-x7g9b\" (UID: \"e87eccc7-4ea1-4cfe-a219-e8d727fd2051\") " pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.071285 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qqwc\" (UniqueName: \"kubernetes.io/projected/ee54b630-23ff-4200-aa84-d3aca72f50e8-kube-api-access-2qqwc\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj8b6\" (UID: \"ee54b630-23ff-4200-aa84-d3aca72f50e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.071337 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xnllr\" (UniqueName: \"kubernetes.io/projected/97cab125-72b2-454c-9bef-bd70f5eb9654-kube-api-access-xnllr\") pod \"package-server-manager-789f6589d5-gwb62\" (UID: \"97cab125-72b2-454c-9bef-bd70f5eb9654\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" Dec 03 17:15:31 crc kubenswrapper[4787]: E1203 17:15:31.075144 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:31.575110892 +0000 UTC m=+148.392581851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.071411 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svrfv\" (UniqueName: \"kubernetes.io/projected/31bdcc6c-aa93-4459-b03f-c559bc726240-kube-api-access-svrfv\") pod \"multus-admission-controller-857f4d67dd-rb22r\" (UID: \"31bdcc6c-aa93-4459-b03f-c559bc726240\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.077107 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-mountpoint-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.077146 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n24hr\" (UniqueName: \"kubernetes.io/projected/440f810f-40c0-4707-afd3-1996592a9cd9-kube-api-access-n24hr\") pod \"migrator-59844c95c7-sl4cp\" (UID: \"440f810f-40c0-4707-afd3-1996592a9cd9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.077195 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e87eccc7-4ea1-4cfe-a219-e8d727fd2051-signing-cabundle\") pod \"service-ca-9c57cc56f-x7g9b\" (UID: \"e87eccc7-4ea1-4cfe-a219-e8d727fd2051\") " pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.077235 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/97cab125-72b2-454c-9bef-bd70f5eb9654-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gwb62\" (UID: \"97cab125-72b2-454c-9bef-bd70f5eb9654\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.077272 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0c6d238-9876-4f67-b557-94ac4f209332-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.077366 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c0c6d238-9876-4f67-b557-94ac4f209332-proxy-tls\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.077425 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-tls\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.077458 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jvwvq\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-kube-api-access-jvwvq\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.082370 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2df4d7cb-f9bb-4ece-9b95-36027a01f017-ca-trust-extracted\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.082822 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-certificates\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.087115 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2df4d7cb-f9bb-4ece-9b95-36027a01f017-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.088422 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/70235f57-caf7-48d6-ab8e-85230e423cd0-service-ca-bundle\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.093993 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-socket-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.094145 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/339b0916-61a9-4e97-9fd3-d3a8dc22d23f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-swngg\" (UID: \"339b0916-61a9-4e97-9fd3-d3a8dc22d23f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.094612 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c09d8632-626c-485d-9e74-65e651546d04-node-bootstrap-token\") pod \"machine-config-server-jjvtn\" (UID: \"c09d8632-626c-485d-9e74-65e651546d04\") " pod="openshift-machine-config-operator/machine-config-server-jjvtn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.096875 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-config-volume\") pod \"collect-profiles-29413035-tghvb\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.097242 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2hvw5\" (UniqueName: \"kubernetes.io/projected/70235f57-caf7-48d6-ab8e-85230e423cd0-kube-api-access-2hvw5\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.097469 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-certificates\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.098330 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/688cf6b7-0d4d-452f-b878-93bd8ff75408-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jgrtv\" (UID: \"688cf6b7-0d4d-452f-b878-93bd8ff75408\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.098385 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2f958667-8c81-40a0-97c0-0afafae369d2-metrics-tls\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.098481 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgrjn\" (UniqueName: \"kubernetes.io/projected/0466f2b9-1263-4369-80f0-364304f1f2f5-kube-api-access-pgrjn\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.099426 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-tls\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.100449 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bl7vk\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.101245 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e87eccc7-4ea1-4cfe-a219-e8d727fd2051-signing-key\") pod \"service-ca-9c57cc56f-x7g9b\" (UID: \"e87eccc7-4ea1-4cfe-a219-e8d727fd2051\") " pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.101380 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.102089 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/552d32f0-c753-4e7d-b0e7-1707e8fa8d26-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gg7bd\" (UID: \"552d32f0-c753-4e7d-b0e7-1707e8fa8d26\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" Dec 03 17:15:31 crc kubenswrapper[4787]: E1203 17:15:31.103483 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:31.603456482 +0000 UTC m=+148.420927441 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.104941 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd4sj\" (UniqueName: \"kubernetes.io/projected/339b0916-61a9-4e97-9fd3-d3a8dc22d23f-kube-api-access-qd4sj\") pod \"kube-storage-version-migrator-operator-b67b599dd-swngg\" (UID: \"339b0916-61a9-4e97-9fd3-d3a8dc22d23f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.104990 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-trusted-ca\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.105228 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aba59ac8-ce2a-4321-a2a3-d790af00541c-config\") pod \"service-ca-operator-777779d784-46g9l\" (UID: \"aba59ac8-ce2a-4321-a2a3-d790af00541c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.105330 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/688cf6b7-0d4d-452f-b878-93bd8ff75408-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jgrtv\" (UID: \"688cf6b7-0d4d-452f-b878-93bd8ff75408\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.105410 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-bound-sa-token\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.105838 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nswv\" (UniqueName: \"kubernetes.io/projected/1af7cc12-ae37-4010-9d9c-28fb02ca03b4-kube-api-access-9nswv\") pod \"catalog-operator-68c6474976-cj96h\" (UID: \"1af7cc12-ae37-4010-9d9c-28fb02ca03b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.105931 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f958667-8c81-40a0-97c0-0afafae369d2-trusted-ca\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.107075 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/31bdcc6c-aa93-4459-b03f-c559bc726240-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rb22r\" (UID: \"31bdcc6c-aa93-4459-b03f-c559bc726240\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.107165 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tm9xq\" (UniqueName: \"kubernetes.io/projected/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-kube-api-access-tm9xq\") pod \"collect-profiles-29413035-tghvb\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.107200 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h7hgx\" (UniqueName: \"kubernetes.io/projected/b3f58901-2824-4993-a0cd-c16e5509894a-kube-api-access-h7hgx\") pod \"dns-default-vjhsg\" (UID: \"b3f58901-2824-4993-a0cd-c16e5509894a\") " pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.107226 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ee54b630-23ff-4200-aa84-d3aca72f50e8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj8b6\" (UID: \"ee54b630-23ff-4200-aa84-d3aca72f50e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.107250 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s69l7\" (UniqueName: \"kubernetes.io/projected/f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7-kube-api-access-s69l7\") pod \"ingress-canary-7m57w\" (UID: \"f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7\") " pod="openshift-ingress-canary/ingress-canary-7m57w" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.107270 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22tgz\" (UniqueName: \"kubernetes.io/projected/c0c6d238-9876-4f67-b557-94ac4f209332-kube-api-access-22tgz\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.107289 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df870037-226f-4286-b128-bbf81e4871b8-proxy-tls\") pod \"machine-config-controller-84d6567774-zk9gm\" (UID: \"df870037-226f-4286-b128-bbf81e4871b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.107372 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/70235f57-caf7-48d6-ab8e-85230e423cd0-default-certificate\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.107401 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552d32f0-c753-4e7d-b0e7-1707e8fa8d26-config\") pod \"kube-controller-manager-operator-78b949d7b-gg7bd\" (UID: \"552d32f0-c753-4e7d-b0e7-1707e8fa8d26\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.107748 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/70235f57-caf7-48d6-ab8e-85230e423cd0-metrics-certs\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.110783 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2f958667-8c81-40a0-97c0-0afafae369d2-trusted-ca\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.120098 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/70235f57-caf7-48d6-ab8e-85230e423cd0-default-certificate\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.127111 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jvwvq\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-kube-api-access-jvwvq\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.130434 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2df4d7cb-f9bb-4ece-9b95-36027a01f017-installation-pull-secrets\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.153956 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2f958667-8c81-40a0-97c0-0afafae369d2-metrics-tls\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.154315 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/70235f57-caf7-48d6-ab8e-85230e423cd0-stats-auth\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.166836 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxlt8\" (UniqueName: \"kubernetes.io/projected/2f958667-8c81-40a0-97c0-0afafae369d2-kube-api-access-xxlt8\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.171510 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2f958667-8c81-40a0-97c0-0afafae369d2-bound-sa-token\") pod \"ingress-operator-5b745b69d9-wb4sr\" (UID: \"2f958667-8c81-40a0-97c0-0afafae369d2\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.197153 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2hvw5\" (UniqueName: \"kubernetes.io/projected/70235f57-caf7-48d6-ab8e-85230e423cd0-kube-api-access-2hvw5\") pod \"router-default-5444994796-x2wsr\" (UID: \"70235f57-caf7-48d6-ab8e-85230e423cd0\") " pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.207989 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.210670 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.210906 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0466f2b9-1263-4369-80f0-364304f1f2f5-apiservice-cert\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.210973 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcnc5\" (UniqueName: \"kubernetes.io/projected/e87eccc7-4ea1-4cfe-a219-e8d727fd2051-kube-api-access-hcnc5\") pod \"service-ca-9c57cc56f-x7g9b\" (UID: \"e87eccc7-4ea1-4cfe-a219-e8d727fd2051\") " pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.210996 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qqwc\" (UniqueName: \"kubernetes.io/projected/ee54b630-23ff-4200-aa84-d3aca72f50e8-kube-api-access-2qqwc\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj8b6\" (UID: \"ee54b630-23ff-4200-aa84-d3aca72f50e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211045 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xnllr\" (UniqueName: \"kubernetes.io/projected/97cab125-72b2-454c-9bef-bd70f5eb9654-kube-api-access-xnllr\") pod \"package-server-manager-789f6589d5-gwb62\" (UID: \"97cab125-72b2-454c-9bef-bd70f5eb9654\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211069 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svrfv\" (UniqueName: \"kubernetes.io/projected/31bdcc6c-aa93-4459-b03f-c559bc726240-kube-api-access-svrfv\") pod \"multus-admission-controller-857f4d67dd-rb22r\" (UID: \"31bdcc6c-aa93-4459-b03f-c559bc726240\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211091 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n24hr\" (UniqueName: \"kubernetes.io/projected/440f810f-40c0-4707-afd3-1996592a9cd9-kube-api-access-n24hr\") pod \"migrator-59844c95c7-sl4cp\" (UID: \"440f810f-40c0-4707-afd3-1996592a9cd9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211114 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-mountpoint-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211137 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e87eccc7-4ea1-4cfe-a219-e8d727fd2051-signing-cabundle\") pod \"service-ca-9c57cc56f-x7g9b\" (UID: \"e87eccc7-4ea1-4cfe-a219-e8d727fd2051\") " pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211161 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/97cab125-72b2-454c-9bef-bd70f5eb9654-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gwb62\" (UID: \"97cab125-72b2-454c-9bef-bd70f5eb9654\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211187 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0c6d238-9876-4f67-b557-94ac4f209332-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211210 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c0c6d238-9876-4f67-b557-94ac4f209332-proxy-tls\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211229 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-socket-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211250 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-config-volume\") pod \"collect-profiles-29413035-tghvb\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211274 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/339b0916-61a9-4e97-9fd3-d3a8dc22d23f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-swngg\" (UID: \"339b0916-61a9-4e97-9fd3-d3a8dc22d23f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211294 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c09d8632-626c-485d-9e74-65e651546d04-node-bootstrap-token\") pod \"machine-config-server-jjvtn\" (UID: \"c09d8632-626c-485d-9e74-65e651546d04\") " pod="openshift-machine-config-operator/machine-config-server-jjvtn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211321 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/688cf6b7-0d4d-452f-b878-93bd8ff75408-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jgrtv\" (UID: \"688cf6b7-0d4d-452f-b878-93bd8ff75408\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211368 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgrjn\" (UniqueName: \"kubernetes.io/projected/0466f2b9-1263-4369-80f0-364304f1f2f5-kube-api-access-pgrjn\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211391 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bl7vk\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211429 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e87eccc7-4ea1-4cfe-a219-e8d727fd2051-signing-key\") pod \"service-ca-9c57cc56f-x7g9b\" (UID: \"e87eccc7-4ea1-4cfe-a219-e8d727fd2051\") " pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211456 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/552d32f0-c753-4e7d-b0e7-1707e8fa8d26-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gg7bd\" (UID: \"552d32f0-c753-4e7d-b0e7-1707e8fa8d26\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211481 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd4sj\" (UniqueName: \"kubernetes.io/projected/339b0916-61a9-4e97-9fd3-d3a8dc22d23f-kube-api-access-qd4sj\") pod \"kube-storage-version-migrator-operator-b67b599dd-swngg\" (UID: \"339b0916-61a9-4e97-9fd3-d3a8dc22d23f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211510 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nswv\" (UniqueName: \"kubernetes.io/projected/1af7cc12-ae37-4010-9d9c-28fb02ca03b4-kube-api-access-9nswv\") pod \"catalog-operator-68c6474976-cj96h\" (UID: \"1af7cc12-ae37-4010-9d9c-28fb02ca03b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211533 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aba59ac8-ce2a-4321-a2a3-d790af00541c-config\") pod \"service-ca-operator-777779d784-46g9l\" (UID: \"aba59ac8-ce2a-4321-a2a3-d790af00541c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211556 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/688cf6b7-0d4d-452f-b878-93bd8ff75408-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jgrtv\" (UID: \"688cf6b7-0d4d-452f-b878-93bd8ff75408\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211577 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/31bdcc6c-aa93-4459-b03f-c559bc726240-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rb22r\" (UID: \"31bdcc6c-aa93-4459-b03f-c559bc726240\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211598 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tm9xq\" (UniqueName: \"kubernetes.io/projected/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-kube-api-access-tm9xq\") pod \"collect-profiles-29413035-tghvb\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211618 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h7hgx\" (UniqueName: \"kubernetes.io/projected/b3f58901-2824-4993-a0cd-c16e5509894a-kube-api-access-h7hgx\") pod \"dns-default-vjhsg\" (UID: \"b3f58901-2824-4993-a0cd-c16e5509894a\") " pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211642 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ee54b630-23ff-4200-aa84-d3aca72f50e8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj8b6\" (UID: \"ee54b630-23ff-4200-aa84-d3aca72f50e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211664 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s69l7\" (UniqueName: \"kubernetes.io/projected/f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7-kube-api-access-s69l7\") pod \"ingress-canary-7m57w\" (UID: \"f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7\") " pod="openshift-ingress-canary/ingress-canary-7m57w" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211687 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22tgz\" (UniqueName: \"kubernetes.io/projected/c0c6d238-9876-4f67-b557-94ac4f209332-kube-api-access-22tgz\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211709 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df870037-226f-4286-b128-bbf81e4871b8-proxy-tls\") pod \"machine-config-controller-84d6567774-zk9gm\" (UID: \"df870037-226f-4286-b128-bbf81e4871b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211733 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552d32f0-c753-4e7d-b0e7-1707e8fa8d26-config\") pod \"kube-controller-manager-operator-78b949d7b-gg7bd\" (UID: \"552d32f0-c753-4e7d-b0e7-1707e8fa8d26\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211756 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c09d8632-626c-485d-9e74-65e651546d04-certs\") pod \"machine-config-server-jjvtn\" (UID: \"c09d8632-626c-485d-9e74-65e651546d04\") " pod="openshift-machine-config-operator/machine-config-server-jjvtn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211779 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-secret-volume\") pod \"collect-profiles-29413035-tghvb\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211802 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f58901-2824-4993-a0cd-c16e5509894a-config-volume\") pod \"dns-default-vjhsg\" (UID: \"b3f58901-2824-4993-a0cd-c16e5509894a\") " pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211826 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88a31888-8dbe-45d6-9ce9-0d9c65e9f914-srv-cert\") pod \"olm-operator-6b444d44fb-zcnjw\" (UID: \"88a31888-8dbe-45d6-9ce9-0d9c65e9f914\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211858 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-csi-data-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211880 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89648a64-0aeb-48b0-ad91-3aac3b4cd5e7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rwvrf\" (UID: \"89648a64-0aeb-48b0-ad91-3aac3b4cd5e7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211900 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-registration-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211921 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b3f58901-2824-4993-a0cd-c16e5509894a-metrics-tls\") pod \"dns-default-vjhsg\" (UID: \"b3f58901-2824-4993-a0cd-c16e5509894a\") " pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211944 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c0c6d238-9876-4f67-b557-94ac4f209332-images\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211968 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ftwj\" (UniqueName: \"kubernetes.io/projected/aba59ac8-ce2a-4321-a2a3-d790af00541c-kube-api-access-5ftwj\") pod \"service-ca-operator-777779d784-46g9l\" (UID: \"aba59ac8-ce2a-4321-a2a3-d790af00541c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.211993 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88a31888-8dbe-45d6-9ce9-0d9c65e9f914-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zcnjw\" (UID: \"88a31888-8dbe-45d6-9ce9-0d9c65e9f914\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212034 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klrdt\" (UniqueName: \"kubernetes.io/projected/c09d8632-626c-485d-9e74-65e651546d04-kube-api-access-klrdt\") pod \"machine-config-server-jjvtn\" (UID: \"c09d8632-626c-485d-9e74-65e651546d04\") " pod="openshift-machine-config-operator/machine-config-server-jjvtn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212054 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb5lp\" (UniqueName: \"kubernetes.io/projected/684133b3-732f-47c6-b15d-108c3e37397f-kube-api-access-mb5lp\") pod \"marketplace-operator-79b997595-bl7vk\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212077 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df870037-226f-4286-b128-bbf81e4871b8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zk9gm\" (UID: \"df870037-226f-4286-b128-bbf81e4871b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212098 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tlzdz\" (UniqueName: \"kubernetes.io/projected/88a31888-8dbe-45d6-9ce9-0d9c65e9f914-kube-api-access-tlzdz\") pod \"olm-operator-6b444d44fb-zcnjw\" (UID: \"88a31888-8dbe-45d6-9ce9-0d9c65e9f914\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212120 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89648a64-0aeb-48b0-ad91-3aac3b4cd5e7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rwvrf\" (UID: \"89648a64-0aeb-48b0-ad91-3aac3b4cd5e7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212136 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/688cf6b7-0d4d-452f-b878-93bd8ff75408-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jgrtv\" (UID: \"688cf6b7-0d4d-452f-b878-93bd8ff75408\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212152 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7-cert\") pod \"ingress-canary-7m57w\" (UID: \"f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7\") " pod="openshift-ingress-canary/ingress-canary-7m57w" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212166 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-plugins-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212181 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65vms\" (UniqueName: \"kubernetes.io/projected/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-kube-api-access-65vms\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212195 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bl7vk\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212214 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cxmp\" (UniqueName: \"kubernetes.io/projected/df870037-226f-4286-b128-bbf81e4871b8-kube-api-access-8cxmp\") pod \"machine-config-controller-84d6567774-zk9gm\" (UID: \"df870037-226f-4286-b128-bbf81e4871b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212232 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1af7cc12-ae37-4010-9d9c-28fb02ca03b4-profile-collector-cert\") pod \"catalog-operator-68c6474976-cj96h\" (UID: \"1af7cc12-ae37-4010-9d9c-28fb02ca03b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212248 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aba59ac8-ce2a-4321-a2a3-d790af00541c-serving-cert\") pod \"service-ca-operator-777779d784-46g9l\" (UID: \"aba59ac8-ce2a-4321-a2a3-d790af00541c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212262 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/552d32f0-c753-4e7d-b0e7-1707e8fa8d26-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gg7bd\" (UID: \"552d32f0-c753-4e7d-b0e7-1707e8fa8d26\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212281 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0466f2b9-1263-4369-80f0-364304f1f2f5-tmpfs\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212300 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0466f2b9-1263-4369-80f0-364304f1f2f5-webhook-cert\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212319 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1af7cc12-ae37-4010-9d9c-28fb02ca03b4-srv-cert\") pod \"catalog-operator-68c6474976-cj96h\" (UID: \"1af7cc12-ae37-4010-9d9c-28fb02ca03b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212333 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/339b0916-61a9-4e97-9fd3-d3a8dc22d23f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-swngg\" (UID: \"339b0916-61a9-4e97-9fd3-d3a8dc22d23f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212351 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4zlf\" (UniqueName: \"kubernetes.io/projected/89648a64-0aeb-48b0-ad91-3aac3b4cd5e7-kube-api-access-b4zlf\") pod \"openshift-apiserver-operator-796bbdcf4f-rwvrf\" (UID: \"89648a64-0aeb-48b0-ad91-3aac3b4cd5e7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.212925 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/0466f2b9-1263-4369-80f0-364304f1f2f5-tmpfs\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.214360 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-config-volume\") pod \"collect-profiles-29413035-tghvb\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:31 crc kubenswrapper[4787]: E1203 17:15:31.214946 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:31.71491215 +0000 UTC m=+148.532383109 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.216567 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.216727 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/339b0916-61a9-4e97-9fd3-d3a8dc22d23f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-swngg\" (UID: \"339b0916-61a9-4e97-9fd3-d3a8dc22d23f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.217793 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-mountpoint-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.222496 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/688cf6b7-0d4d-452f-b878-93bd8ff75408-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jgrtv\" (UID: \"688cf6b7-0d4d-452f-b878-93bd8ff75408\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.232238 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/aba59ac8-ce2a-4321-a2a3-d790af00541c-config\") pod \"service-ca-operator-777779d784-46g9l\" (UID: \"aba59ac8-ce2a-4321-a2a3-d790af00541c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.232374 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/df870037-226f-4286-b128-bbf81e4871b8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-zk9gm\" (UID: \"df870037-226f-4286-b128-bbf81e4871b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.232473 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/e87eccc7-4ea1-4cfe-a219-e8d727fd2051-signing-cabundle\") pod \"service-ca-9c57cc56f-x7g9b\" (UID: \"e87eccc7-4ea1-4cfe-a219-e8d727fd2051\") " pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.233402 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bl7vk\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.235190 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/552d32f0-c753-4e7d-b0e7-1707e8fa8d26-config\") pod \"kube-controller-manager-operator-78b949d7b-gg7bd\" (UID: \"552d32f0-c753-4e7d-b0e7-1707e8fa8d26\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.235309 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b3f58901-2824-4993-a0cd-c16e5509894a-config-volume\") pod \"dns-default-vjhsg\" (UID: \"b3f58901-2824-4993-a0cd-c16e5509894a\") " pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.235401 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-plugins-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.235622 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-socket-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.237761 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/89648a64-0aeb-48b0-ad91-3aac3b4cd5e7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-rwvrf\" (UID: \"89648a64-0aeb-48b0-ad91-3aac3b4cd5e7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.240087 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-csi-data-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.240094 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-registration-dir\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.241903 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/c0c6d238-9876-4f67-b557-94ac4f209332-images\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.245187 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/0466f2b9-1263-4369-80f0-364304f1f2f5-apiservice-cert\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.245221 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bl7vk\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.247212 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b3f58901-2824-4993-a0cd-c16e5509894a-metrics-tls\") pod \"dns-default-vjhsg\" (UID: \"b3f58901-2824-4993-a0cd-c16e5509894a\") " pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.252331 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/688cf6b7-0d4d-452f-b878-93bd8ff75408-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jgrtv\" (UID: \"688cf6b7-0d4d-452f-b878-93bd8ff75408\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.253005 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/ee54b630-23ff-4200-aa84-d3aca72f50e8-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj8b6\" (UID: \"ee54b630-23ff-4200-aa84-d3aca72f50e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.253468 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/c0c6d238-9876-4f67-b557-94ac4f209332-auth-proxy-config\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.253559 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/1af7cc12-ae37-4010-9d9c-28fb02ca03b4-srv-cert\") pod \"catalog-operator-68c6474976-cj96h\" (UID: \"1af7cc12-ae37-4010-9d9c-28fb02ca03b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.254113 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/88a31888-8dbe-45d6-9ce9-0d9c65e9f914-profile-collector-cert\") pod \"olm-operator-6b444d44fb-zcnjw\" (UID: \"88a31888-8dbe-45d6-9ce9-0d9c65e9f914\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.254352 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/0466f2b9-1263-4369-80f0-364304f1f2f5-webhook-cert\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.254577 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/88a31888-8dbe-45d6-9ce9-0d9c65e9f914-srv-cert\") pod \"olm-operator-6b444d44fb-zcnjw\" (UID: \"88a31888-8dbe-45d6-9ce9-0d9c65e9f914\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.254918 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/31bdcc6c-aa93-4459-b03f-c559bc726240-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-rb22r\" (UID: \"31bdcc6c-aa93-4459-b03f-c559bc726240\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.258462 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/97cab125-72b2-454c-9bef-bd70f5eb9654-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-gwb62\" (UID: \"97cab125-72b2-454c-9bef-bd70f5eb9654\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.259595 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/c09d8632-626c-485d-9e74-65e651546d04-node-bootstrap-token\") pod \"machine-config-server-jjvtn\" (UID: \"c09d8632-626c-485d-9e74-65e651546d04\") " pod="openshift-machine-config-operator/machine-config-server-jjvtn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.260324 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/c09d8632-626c-485d-9e74-65e651546d04-certs\") pod \"machine-config-server-jjvtn\" (UID: \"c09d8632-626c-485d-9e74-65e651546d04\") " pod="openshift-machine-config-operator/machine-config-server-jjvtn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.260680 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/c0c6d238-9876-4f67-b557-94ac4f209332-proxy-tls\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.260897 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/552d32f0-c753-4e7d-b0e7-1707e8fa8d26-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-gg7bd\" (UID: \"552d32f0-c753-4e7d-b0e7-1707e8fa8d26\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.260933 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/e87eccc7-4ea1-4cfe-a219-e8d727fd2051-signing-key\") pod \"service-ca-9c57cc56f-x7g9b\" (UID: \"e87eccc7-4ea1-4cfe-a219-e8d727fd2051\") " pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.262075 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7-cert\") pod \"ingress-canary-7m57w\" (UID: \"f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7\") " pod="openshift-ingress-canary/ingress-canary-7m57w" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.262142 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/df870037-226f-4286-b128-bbf81e4871b8-proxy-tls\") pod \"machine-config-controller-84d6567774-zk9gm\" (UID: \"df870037-226f-4286-b128-bbf81e4871b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.262721 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/1af7cc12-ae37-4010-9d9c-28fb02ca03b4-profile-collector-cert\") pod \"catalog-operator-68c6474976-cj96h\" (UID: \"1af7cc12-ae37-4010-9d9c-28fb02ca03b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.268612 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/339b0916-61a9-4e97-9fd3-d3a8dc22d23f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-swngg\" (UID: \"339b0916-61a9-4e97-9fd3-d3a8dc22d23f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.269542 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/89648a64-0aeb-48b0-ad91-3aac3b4cd5e7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-rwvrf\" (UID: \"89648a64-0aeb-48b0-ad91-3aac3b4cd5e7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.270892 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/aba59ac8-ce2a-4321-a2a3-d790af00541c-serving-cert\") pod \"service-ca-operator-777779d784-46g9l\" (UID: \"aba59ac8-ce2a-4321-a2a3-d790af00541c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.272603 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ftwj\" (UniqueName: \"kubernetes.io/projected/aba59ac8-ce2a-4321-a2a3-d790af00541c-kube-api-access-5ftwj\") pod \"service-ca-operator-777779d784-46g9l\" (UID: \"aba59ac8-ce2a-4321-a2a3-d790af00541c\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.275843 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-bound-sa-token\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.276718 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-secret-volume\") pod \"collect-profiles-29413035-tghvb\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.296618 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.300301 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s69l7\" (UniqueName: \"kubernetes.io/projected/f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7-kube-api-access-s69l7\") pod \"ingress-canary-7m57w\" (UID: \"f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7\") " pod="openshift-ingress-canary/ingress-canary-7m57w" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.313564 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4zlf\" (UniqueName: \"kubernetes.io/projected/89648a64-0aeb-48b0-ad91-3aac3b4cd5e7-kube-api-access-b4zlf\") pod \"openshift-apiserver-operator-796bbdcf4f-rwvrf\" (UID: \"89648a64-0aeb-48b0-ad91-3aac3b4cd5e7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.313957 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: E1203 17:15:31.314290 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:31.814277819 +0000 UTC m=+148.631748778 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:31 crc kubenswrapper[4787]: W1203 17:15:31.372176 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod70235f57_caf7_48d6_ab8e_85230e423cd0.slice/crio-a30071d9592542d813d56e031df84db46766a7d45f56e1f0f714fcf2f74c57f3 WatchSource:0}: Error finding container a30071d9592542d813d56e031df84db46766a7d45f56e1f0f714fcf2f74c57f3: Status 404 returned error can't find the container with id a30071d9592542d813d56e031df84db46766a7d45f56e1f0f714fcf2f74c57f3 Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.373349 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.387037 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22tgz\" (UniqueName: \"kubernetes.io/projected/c0c6d238-9876-4f67-b557-94ac4f209332-kube-api-access-22tgz\") pod \"machine-config-operator-74547568cd-d72l6\" (UID: \"c0c6d238-9876-4f67-b557-94ac4f209332\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.388902 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/552d32f0-c753-4e7d-b0e7-1707e8fa8d26-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-gg7bd\" (UID: \"552d32f0-c753-4e7d-b0e7-1707e8fa8d26\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.391734 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-7m57w" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.414644 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:31 crc kubenswrapper[4787]: E1203 17:15:31.414790 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:31.914766058 +0000 UTC m=+148.732237017 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.415221 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: E1203 17:15:31.415659 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:31.915641712 +0000 UTC m=+148.733112671 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.441598 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xnllr\" (UniqueName: \"kubernetes.io/projected/97cab125-72b2-454c-9bef-bd70f5eb9654-kube-api-access-xnllr\") pod \"package-server-manager-789f6589d5-gwb62\" (UID: \"97cab125-72b2-454c-9bef-bd70f5eb9654\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.443416 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svrfv\" (UniqueName: \"kubernetes.io/projected/31bdcc6c-aa93-4459-b03f-c559bc726240-kube-api-access-svrfv\") pod \"multus-admission-controller-857f4d67dd-rb22r\" (UID: \"31bdcc6c-aa93-4459-b03f-c559bc726240\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.444950 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qqwc\" (UniqueName: \"kubernetes.io/projected/ee54b630-23ff-4200-aa84-d3aca72f50e8-kube-api-access-2qqwc\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj8b6\" (UID: \"ee54b630-23ff-4200-aa84-d3aca72f50e8\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.449690 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcnc5\" (UniqueName: \"kubernetes.io/projected/e87eccc7-4ea1-4cfe-a219-e8d727fd2051-kube-api-access-hcnc5\") pod \"service-ca-9c57cc56f-x7g9b\" (UID: \"e87eccc7-4ea1-4cfe-a219-e8d727fd2051\") " pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.463766 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n24hr\" (UniqueName: \"kubernetes.io/projected/440f810f-40c0-4707-afd3-1996592a9cd9-kube-api-access-n24hr\") pod \"migrator-59844c95c7-sl4cp\" (UID: \"440f810f-40c0-4707-afd3-1996592a9cd9\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.467707 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd4sj\" (UniqueName: \"kubernetes.io/projected/339b0916-61a9-4e97-9fd3-d3a8dc22d23f-kube-api-access-qd4sj\") pod \"kube-storage-version-migrator-operator-b67b599dd-swngg\" (UID: \"339b0916-61a9-4e97-9fd3-d3a8dc22d23f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.479256 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tm9xq\" (UniqueName: \"kubernetes.io/projected/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-kube-api-access-tm9xq\") pod \"collect-profiles-29413035-tghvb\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.502664 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm"] Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.511435 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nswv\" (UniqueName: \"kubernetes.io/projected/1af7cc12-ae37-4010-9d9c-28fb02ca03b4-kube-api-access-9nswv\") pod \"catalog-operator-68c6474976-cj96h\" (UID: \"1af7cc12-ae37-4010-9d9c-28fb02ca03b4\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.512392 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-f2vjk"] Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.519205 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:31 crc kubenswrapper[4787]: E1203 17:15:31.519619 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.019600716 +0000 UTC m=+148.837071675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.529801 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klrdt\" (UniqueName: \"kubernetes.io/projected/c09d8632-626c-485d-9e74-65e651546d04-kube-api-access-klrdt\") pod \"machine-config-server-jjvtn\" (UID: \"c09d8632-626c-485d-9e74-65e651546d04\") " pod="openshift-machine-config-operator/machine-config-server-jjvtn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.543602 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb5lp\" (UniqueName: \"kubernetes.io/projected/684133b3-732f-47c6-b15d-108c3e37397f-kube-api-access-mb5lp\") pod \"marketplace-operator-79b997595-bl7vk\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.545583 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-gdftf"] Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.557678 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65vms\" (UniqueName: \"kubernetes.io/projected/f2d70e9f-74fc-47ca-9e05-4daf81c06c7b-kube-api-access-65vms\") pod \"csi-hostpathplugin-cggxn\" (UID: \"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b\") " pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.574147 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tlzdz\" (UniqueName: \"kubernetes.io/projected/88a31888-8dbe-45d6-9ce9-0d9c65e9f914-kube-api-access-tlzdz\") pod \"olm-operator-6b444d44fb-zcnjw\" (UID: \"88a31888-8dbe-45d6-9ce9-0d9c65e9f914\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.588850 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.595357 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cxmp\" (UniqueName: \"kubernetes.io/projected/df870037-226f-4286-b128-bbf81e4871b8-kube-api-access-8cxmp\") pod \"machine-config-controller-84d6567774-zk9gm\" (UID: \"df870037-226f-4286-b128-bbf81e4871b8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.605277 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6" Dec 03 17:15:31 crc kubenswrapper[4787]: W1203 17:15:31.608381 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf637fc79_77c3_492a_8d39_c5c8454fd322.slice/crio-6bcff3db565662bb39260df3b494a8ce04dfa48c72b9804ffeb55820e8745828 WatchSource:0}: Error finding container 6bcff3db565662bb39260df3b494a8ce04dfa48c72b9804ffeb55820e8745828: Status 404 returned error can't find the container with id 6bcff3db565662bb39260df3b494a8ce04dfa48c72b9804ffeb55820e8745828 Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.611265 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.613809 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgrjn\" (UniqueName: \"kubernetes.io/projected/0466f2b9-1263-4369-80f0-364304f1f2f5-kube-api-access-pgrjn\") pod \"packageserver-d55dfcdfc-h5zms\" (UID: \"0466f2b9-1263-4369-80f0-364304f1f2f5\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.625590 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.626328 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" Dec 03 17:15:31 crc kubenswrapper[4787]: E1203 17:15:31.626367 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.126347955 +0000 UTC m=+148.943818914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.637196 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.640625 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.649364 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.654926 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h7hgx\" (UniqueName: \"kubernetes.io/projected/b3f58901-2824-4993-a0cd-c16e5509894a-kube-api-access-h7hgx\") pod \"dns-default-vjhsg\" (UID: \"b3f58901-2824-4993-a0cd-c16e5509894a\") " pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.656349 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.660013 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-pjpcp" event={"ID":"844a7007-5cb5-42b1-ada8-9b4faf3dddc1","Type":"ContainerStarted","Data":"42e10d255ee34a5eff98e657c0b38b3e83d517d405cd44a4157643b42a0d1ce8"} Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.661563 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.661988 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.673965 4787 patch_prober.go:28] interesting pod/console-operator-58897d9998-pjpcp container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.674080 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-pjpcp" podUID="844a7007-5cb5-42b1-ada8-9b4faf3dddc1" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/readyz\": dial tcp 10.217.0.20:8443: connect: connection refused" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.675408 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.675562 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/688cf6b7-0d4d-452f-b878-93bd8ff75408-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-jgrtv\" (UID: \"688cf6b7-0d4d-452f-b878-93bd8ff75408\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.676075 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" event={"ID":"f637fc79-77c3-492a-8d39-c5c8454fd322","Type":"ContainerStarted","Data":"6bcff3db565662bb39260df3b494a8ce04dfa48c72b9804ffeb55820e8745828"} Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.676631 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.687393 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.699881 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-jjvtn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.703165 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" event={"ID":"48860c61-124e-43cf-9cf6-fd36f33866f5","Type":"ContainerStarted","Data":"6fe02eb058e3fe80e67e47b0f839e8dbaeea413440f921059ff82a06c99d851f"} Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.705492 4787 generic.go:334] "Generic (PLEG): container finished" podID="8df314ae-77d5-48cb-9d05-6737236e0a9f" containerID="a3c0ff15cda9d2547c1c809fc02afdbc9d921892ff5ff5fa04a3cba8812bb567" exitCode=0 Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.705550 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" event={"ID":"8df314ae-77d5-48cb-9d05-6737236e0a9f","Type":"ContainerDied","Data":"a3c0ff15cda9d2547c1c809fc02afdbc9d921892ff5ff5fa04a3cba8812bb567"} Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.706197 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.730647 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:31 crc kubenswrapper[4787]: E1203 17:15:31.735426 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.235396688 +0000 UTC m=+149.052867647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.743460 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-cggxn" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.799003 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-x2wsr" event={"ID":"70235f57-caf7-48d6-ab8e-85230e423cd0","Type":"ContainerStarted","Data":"a30071d9592542d813d56e031df84db46766a7d45f56e1f0f714fcf2f74c57f3"} Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.811855 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" event={"ID":"ea7b4185-4d6b-4d35-af22-252dc98d140d","Type":"ContainerStarted","Data":"2f30b61f2367787df71981482c5d6d57072e5ccc6216e8b9c4985fc6c9b3cc88"} Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.833337 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:31 crc kubenswrapper[4787]: E1203 17:15:31.835053 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.335041854 +0000 UTC m=+149.152512803 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.855045 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" event={"ID":"27bf1dec-a834-41e8-acdd-e5047002ed0b","Type":"ContainerStarted","Data":"118795c4f5b6f2e6d212de01aa3d145359f2042795265d42eefb66fd0b639d18"} Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.867326 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.875711 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.885376 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:31 crc kubenswrapper[4787]: I1203 17:15:31.969087 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:31 crc kubenswrapper[4787]: E1203 17:15:31.971081 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.471032828 +0000 UTC m=+149.288503777 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:31.992818 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" event={"ID":"e19b19ce-efa6-4d43-af6c-bb6aed7c232d","Type":"ContainerStarted","Data":"24cf7301af91405740dd42a3d2da7c37a452748aed65f4f10b6c03138839da9b"} Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:31.994125 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.028178 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" event={"ID":"0cc4a1a3-bd7c-4ed7-a818-a1cef15dce4e","Type":"ContainerStarted","Data":"f4653b33686d79502a5f1541faf7d04e22a794371a97293122bfaeb17bd98c91"} Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.081744 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.082137 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.582108114 +0000 UTC m=+149.399579073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.159543 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" event={"ID":"9d570d04-9282-45fd-848b-7752fa9062a0","Type":"ContainerStarted","Data":"421ba62825562d2d4a1220007eec5a2ae4d4b4433233d1a3d072a46d00ea8355"} Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.182584 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.184387 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.684364502 +0000 UTC m=+149.501835461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.204568 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" event={"ID":"025afbb8-3e97-4e1e-a1c8-c50af81daba2","Type":"ContainerStarted","Data":"119ad0028f1dbd22bea67ecc2ce55c97d5361601078f1602c1e1b2417d88ddd0"} Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.232111 4787 generic.go:334] "Generic (PLEG): container finished" podID="1c90cc96-7e0f-445a-9519-f1632e4b0b5d" containerID="06a41cc4fb2c9211b5b24607463d59f3ae2ec9786149d1a424864092774d6798" exitCode=0 Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.232217 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" event={"ID":"1c90cc96-7e0f-445a-9519-f1632e4b0b5d","Type":"ContainerDied","Data":"06a41cc4fb2c9211b5b24607463d59f3ae2ec9786149d1a424864092774d6798"} Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.270635 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" event={"ID":"670e0167-14c0-4547-9539-f303e18c33f3","Type":"ContainerStarted","Data":"f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431"} Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.270704 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" event={"ID":"670e0167-14c0-4547-9539-f303e18c33f3","Type":"ContainerStarted","Data":"23ee204b10f5e8af6ca66522a09e820fb3039e17b6c10321e50a807949808c64"} Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.272286 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.285474 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.286110 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.786088955 +0000 UTC m=+149.603559914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.315340 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" event={"ID":"12108aae-ded8-4fc7-b763-bd6bfd13aa81","Type":"ContainerStarted","Data":"a5fa2f1d7c07526a70073e2bf88d0d70d5f6a40db20e67fe1ded05c26aa7f50f"} Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.334772 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" event={"ID":"4cf3d039-080b-4a38-a7e5-0a1e02f9f526","Type":"ContainerStarted","Data":"2eb7a8f88d38f5843d2091f20c1582a3abc7629d74d272479dccf450483d07f4"} Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.342265 4787 patch_prober.go:28] interesting pod/downloads-7954f5f757-xft52 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.342337 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xft52" podUID="ade4b5ab-2584-49e9-9ece-a02a16d79a99" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.344943 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-46g9l"] Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.367732 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.386336 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.386756 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.886707518 +0000 UTC m=+149.704178477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.387265 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.388988 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.888974139 +0000 UTC m=+149.706445098 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.389899 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf"] Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.396680 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-7m57w"] Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.430469 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr"] Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.433432 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.488608 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.488746 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.988723909 +0000 UTC m=+149.806194878 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.490140 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.491494 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:32.991474194 +0000 UTC m=+149.808945153 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.597906 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.601588 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:33.101551534 +0000 UTC m=+149.919022493 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.676169 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.706687 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.707307 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:33.207278876 +0000 UTC m=+150.024749845 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.819519 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:33.319482753 +0000 UTC m=+150.136953712 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.819564 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.820051 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.820584 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:33.320577743 +0000 UTC m=+150.138048692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.917707 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6"] Dec 03 17:15:32 crc kubenswrapper[4787]: I1203 17:15:32.922343 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:32 crc kubenswrapper[4787]: E1203 17:15:32.923509 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:33.423471868 +0000 UTC m=+150.240942827 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.026080 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.026683 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:33.526650521 +0000 UTC m=+150.344121470 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.131524 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.131753 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:33.631715705 +0000 UTC m=+150.449186664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.132045 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.132404 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:33.632396233 +0000 UTC m=+150.449867192 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: W1203 17:15:33.228087 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc09d8632_626c_485d_9e74_65e651546d04.slice/crio-4e36cbfd0b11389969f31163e1095c0dfb22f563ea30abf17d2e1cc8f20855c5 WatchSource:0}: Error finding container 4e36cbfd0b11389969f31163e1095c0dfb22f563ea30abf17d2e1cc8f20855c5: Status 404 returned error can't find the container with id 4e36cbfd0b11389969f31163e1095c0dfb22f563ea30abf17d2e1cc8f20855c5 Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.236006 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.236271 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:33.736256364 +0000 UTC m=+150.553727323 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.285898 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h"] Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.307669 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bl7vk"] Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.342593 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.343899 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:33.843880718 +0000 UTC m=+150.661351677 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.443093 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" event={"ID":"2f958667-8c81-40a0-97c0-0afafae369d2","Type":"ContainerStarted","Data":"8eb4b3520d531586d4c27120f4368e49a80f6d4a9d89c68d1a3d322ee4669895"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.445795 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.446135 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:33.946111655 +0000 UTC m=+150.763582604 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.482612 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7m57w" event={"ID":"f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7","Type":"ContainerStarted","Data":"372d72a2a104a8afbfb280e30f5a868ca3381b739e2d10bce9b90a41bf29143b"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.495402 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" event={"ID":"89648a64-0aeb-48b0-ad91-3aac3b4cd5e7","Type":"ContainerStarted","Data":"509463da33b90ddb0eaf3952e3a7b77d824b516d0a6d2a643658832bdd92f075"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.547276 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.547881 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:34.047867999 +0000 UTC m=+150.865338958 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.590771 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-l4sbk" podStartSLOduration=131.590732863 podStartE2EDuration="2m11.590732863s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:33.529728446 +0000 UTC m=+150.347199405" watchObservedRunningTime="2025-12-03 17:15:33.590732863 +0000 UTC m=+150.408203822" Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.600231 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" event={"ID":"aba59ac8-ce2a-4321-a2a3-d790af00541c","Type":"ContainerStarted","Data":"4290463c4b952896f0a8f69699e393e8209e324409b21f88813a31f8b1816775"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.639689 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-jjvtn" event={"ID":"c09d8632-626c-485d-9e74-65e651546d04","Type":"ContainerStarted","Data":"4e36cbfd0b11389969f31163e1095c0dfb22f563ea30abf17d2e1cc8f20855c5"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.642671 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"13e2f4981cc263cc7f5355e76eb253ef96da414cc656d5f07077166abc22c66f"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.642973 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" podStartSLOduration=131.642944461 podStartE2EDuration="2m11.642944461s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:33.640810603 +0000 UTC m=+150.458281572" watchObservedRunningTime="2025-12-03 17:15:33.642944461 +0000 UTC m=+150.460415420" Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.643363 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6" event={"ID":"ee54b630-23ff-4200-aa84-d3aca72f50e8","Type":"ContainerStarted","Data":"11d696f0cbd662f9a7108508c69316f59823e292008cbd3ef3caff3c649e16db"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.643912 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-p875r" podStartSLOduration=131.643902917 podStartE2EDuration="2m11.643902917s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:33.596720406 +0000 UTC m=+150.414191365" watchObservedRunningTime="2025-12-03 17:15:33.643902917 +0000 UTC m=+150.461373876" Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.648468 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-x2wsr" event={"ID":"70235f57-caf7-48d6-ab8e-85230e423cd0","Type":"ContainerStarted","Data":"e33d8c9438cf3ff8c67038dd881b2f4875755d067c39b83a5e5564adf7db4909"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.649274 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.649628 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:34.149606832 +0000 UTC m=+150.967077791 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.676838 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"33d210b022d851b3ffb3e4f8562977b7d5467f05ac9dc90efd5995ae3249568f"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.678252 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d29c2d4c1673bfab1d117c83040f30f80c76a682675d1a77a9c7e80c1fc96501"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.682575 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" event={"ID":"9d570d04-9282-45fd-848b-7752fa9062a0","Type":"ContainerStarted","Data":"b7faf703b4a8a0bc8ca738aa76b3a5e980cc46fcb33341da7f53102644f3a4b6"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.701473 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-sm94s" podStartSLOduration=131.70145489 podStartE2EDuration="2m11.70145489s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:33.700271208 +0000 UTC m=+150.517742177" watchObservedRunningTime="2025-12-03 17:15:33.70145489 +0000 UTC m=+150.518925849" Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.732212 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" event={"ID":"27bf1dec-a834-41e8-acdd-e5047002ed0b","Type":"ContainerStarted","Data":"3acdefbb9211238c8c70f879dfb6c4b1f5b66ac8ccee1db934367982d642322a"} Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.755163 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.755583 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:34.25556785 +0000 UTC m=+151.073038809 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.767178 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-pjpcp" Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.775069 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-xft52" podStartSLOduration=131.775046859 podStartE2EDuration="2m11.775046859s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:33.758719226 +0000 UTC m=+150.576190185" watchObservedRunningTime="2025-12-03 17:15:33.775046859 +0000 UTC m=+150.592517818" Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.856253 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-pjpcp" podStartSLOduration=131.856228255 podStartE2EDuration="2m11.856228255s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:33.814346417 +0000 UTC m=+150.631817376" watchObservedRunningTime="2025-12-03 17:15:33.856228255 +0000 UTC m=+150.673699214" Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.857859 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb"] Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.859062 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.859512 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:34.359476353 +0000 UTC m=+151.176947312 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.859901 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.866612 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:34.366586346 +0000 UTC m=+151.184057305 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.894460 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" podStartSLOduration=131.894445673 podStartE2EDuration="2m11.894445673s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:33.892651814 +0000 UTC m=+150.710122763" watchObservedRunningTime="2025-12-03 17:15:33.894445673 +0000 UTC m=+150.711916632" Dec 03 17:15:33 crc kubenswrapper[4787]: I1203 17:15:33.960883 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:33 crc kubenswrapper[4787]: E1203 17:15:33.961181 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:34.461159645 +0000 UTC m=+151.278630604 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.022494 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-rhg28" podStartSLOduration=133.02247271 podStartE2EDuration="2m13.02247271s" podCreationTimestamp="2025-12-03 17:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:34.005073097 +0000 UTC m=+150.822544066" watchObservedRunningTime="2025-12-03 17:15:34.02247271 +0000 UTC m=+150.839943679" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.060319 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rcs6j" podStartSLOduration=132.060279677 podStartE2EDuration="2m12.060279677s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:34.059007342 +0000 UTC m=+150.876478311" watchObservedRunningTime="2025-12-03 17:15:34.060279677 +0000 UTC m=+150.877750636" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.063590 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:34 crc kubenswrapper[4787]: E1203 17:15:34.063950 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:34.563937386 +0000 UTC m=+151.381408345 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.100661 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-nm7m8" podStartSLOduration=132.100645794 podStartE2EDuration="2m12.100645794s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:34.09792985 +0000 UTC m=+150.915400799" watchObservedRunningTime="2025-12-03 17:15:34.100645794 +0000 UTC m=+150.918116753" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.164397 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:34 crc kubenswrapper[4787]: E1203 17:15:34.164514 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:34.664496298 +0000 UTC m=+151.481967257 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.164621 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:34 crc kubenswrapper[4787]: E1203 17:15:34.164917 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:34.664908529 +0000 UTC m=+151.482379488 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.188885 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-86q7g" podStartSLOduration=133.18886326 podStartE2EDuration="2m13.18886326s" podCreationTimestamp="2025-12-03 17:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:34.143314073 +0000 UTC m=+150.960785032" watchObservedRunningTime="2025-12-03 17:15:34.18886326 +0000 UTC m=+151.006334219" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.220312 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.253293 4787 patch_prober.go:28] interesting pod/router-default-5444994796-x2wsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 17:15:34 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Dec 03 17:15:34 crc kubenswrapper[4787]: [+]process-running ok Dec 03 17:15:34 crc kubenswrapper[4787]: healthz check failed Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.253542 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x2wsr" podUID="70235f57-caf7-48d6-ab8e-85230e423cd0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.292257 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:34 crc kubenswrapper[4787]: E1203 17:15:34.293233 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:34.793165433 +0000 UTC m=+151.610636392 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.301660 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" podStartSLOduration=132.301614062 podStartE2EDuration="2m12.301614062s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:34.269806188 +0000 UTC m=+151.087277147" watchObservedRunningTime="2025-12-03 17:15:34.301614062 +0000 UTC m=+151.119085021" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.394493 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:34 crc kubenswrapper[4787]: E1203 17:15:34.405587 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:34.905562216 +0000 UTC m=+151.723033175 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.406237 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vjhsg"] Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.406240 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-x2wsr" podStartSLOduration=132.406215634 podStartE2EDuration="2m12.406215634s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:34.400820487 +0000 UTC m=+151.218291446" watchObservedRunningTime="2025-12-03 17:15:34.406215634 +0000 UTC m=+151.223686593" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.462914 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" podStartSLOduration=133.462883203 podStartE2EDuration="2m13.462883203s" podCreationTimestamp="2025-12-03 17:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:34.45392944 +0000 UTC m=+151.271400399" watchObservedRunningTime="2025-12-03 17:15:34.462883203 +0000 UTC m=+151.280354162" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.470305 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv"] Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.523320 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:34 crc kubenswrapper[4787]: E1203 17:15:34.523798 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:35.023778807 +0000 UTC m=+151.841249766 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.567189 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-rb22r"] Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.624883 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:34 crc kubenswrapper[4787]: E1203 17:15:34.625322 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:35.125305855 +0000 UTC m=+151.942776814 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.668927 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd"] Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.710959 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6"] Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.724681 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.725187 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.725893 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:34 crc kubenswrapper[4787]: E1203 17:15:34.726507 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:35.226483623 +0000 UTC m=+152.043954582 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.774628 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-7m57w" event={"ID":"f1f31e68-d5a7-4e13-a88c-46e2cbfe55f7","Type":"ContainerStarted","Data":"04b1d300cb92b3cdd013b4586a3809b5a552bb2bf1a5a570f4f0df63f5044057"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.784581 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" event={"ID":"89648a64-0aeb-48b0-ad91-3aac3b4cd5e7","Type":"ContainerStarted","Data":"02c50bea7c19de0ef0a625b7e1ae1fa70296bf1c1da9dbb42a380f292092bb2a"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.805402 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw"] Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.826314 4787 patch_prober.go:28] interesting pod/apiserver-76f77b778f-lwzb2 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 03 17:15:34 crc kubenswrapper[4787]: [+]log ok Dec 03 17:15:34 crc kubenswrapper[4787]: [+]etcd ok Dec 03 17:15:34 crc kubenswrapper[4787]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 03 17:15:34 crc kubenswrapper[4787]: [+]poststarthook/generic-apiserver-start-informers ok Dec 03 17:15:34 crc kubenswrapper[4787]: [+]poststarthook/max-in-flight-filter ok Dec 03 17:15:34 crc kubenswrapper[4787]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 03 17:15:34 crc kubenswrapper[4787]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 03 17:15:34 crc kubenswrapper[4787]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 03 17:15:34 crc kubenswrapper[4787]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 03 17:15:34 crc kubenswrapper[4787]: [+]poststarthook/project.openshift.io-projectcache ok Dec 03 17:15:34 crc kubenswrapper[4787]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 03 17:15:34 crc kubenswrapper[4787]: [+]poststarthook/openshift.io-startinformers ok Dec 03 17:15:34 crc kubenswrapper[4787]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 03 17:15:34 crc kubenswrapper[4787]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 03 17:15:34 crc kubenswrapper[4787]: livez check failed Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.826379 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" podUID="9d570d04-9282-45fd-848b-7752fa9062a0" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.826551 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a0acb896953dc9c699a2fd27e9b48b58e8ba0a86e609a5bf36f6f708e7ad2603"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.827459 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:34 crc kubenswrapper[4787]: E1203 17:15:34.827930 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:35.327914138 +0000 UTC m=+152.145385107 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.838562 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" event={"ID":"f637fc79-77c3-492a-8d39-c5c8454fd322","Type":"ContainerStarted","Data":"cd50a22623e71231bd63f203bd95a97a134bf9c669c35d0a53a4e62441d1d3e0"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.859707 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" event={"ID":"684133b3-732f-47c6-b15d-108c3e37397f","Type":"ContainerStarted","Data":"dd3a47df7001bc1a54a3c7851801c012a5d79a8cb331912d4e841556d4110921"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.882061 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" event={"ID":"12108aae-ded8-4fc7-b763-bd6bfd13aa81","Type":"ContainerStarted","Data":"dac25147600acaa820dd88306a92a9fb15d527dfe8e8cc37b702bb9af38b284c"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.891715 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" event={"ID":"8df314ae-77d5-48cb-9d05-6737236e0a9f","Type":"ContainerStarted","Data":"fca42928ed35db2909271052bfd2d6c4f370d5de7136b6165fa39dcbd5e21e55"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.895161 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.897914 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"034d8c9640cfd3fee32cba02aeae1ea2526d83de3109f8625f718e5e045d7302"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.913179 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" event={"ID":"1af7cc12-ae37-4010-9d9c-28fb02ca03b4","Type":"ContainerStarted","Data":"ff0af123bd4e91137a1ef623ae6ddcdc063eaebf0048e68c7222329799d1f446"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.914108 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vjhsg" event={"ID":"b3f58901-2824-4993-a0cd-c16e5509894a","Type":"ContainerStarted","Data":"5f43d62eafa22331a5560de5cab24a5618c520675192b1f9ba2f8873565e75f3"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.915295 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ee1e298fa568abfb2e91bb54ba72e8ffe19c3d6ac06a91352def2e1ac0170aa7"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.917290 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.936901 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:34 crc kubenswrapper[4787]: E1203 17:15:34.938641 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:35.438619376 +0000 UTC m=+152.256090335 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.940165 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-cggxn"] Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.951907 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" event={"ID":"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75","Type":"ContainerStarted","Data":"621c87e9f8d3a5756d315d50e416fe9a4032e60dd1b229edac51a1b30032e02c"} Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.971501 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg"] Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.996167 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp"] Dec 03 17:15:34 crc kubenswrapper[4787]: I1203 17:15:34.999922 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" event={"ID":"1c90cc96-7e0f-445a-9519-f1632e4b0b5d","Type":"ContainerStarted","Data":"08b8a4682e170bf636892cc3dc769ffcf20c0de3bf7938f8871b389afaa7842e"} Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.045165 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:35 crc kubenswrapper[4787]: E1203 17:15:35.070496 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:35.570480027 +0000 UTC m=+152.387950986 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.147338 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:35 crc kubenswrapper[4787]: E1203 17:15:35.147621 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:35.647598652 +0000 UTC m=+152.465069611 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.174500 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-7m57w" podStartSLOduration=7.174484122 podStartE2EDuration="7.174484122s" podCreationTimestamp="2025-12-03 17:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:35.172724575 +0000 UTC m=+151.990195534" watchObservedRunningTime="2025-12-03 17:15:35.174484122 +0000 UTC m=+151.991955081" Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.234443 4787 patch_prober.go:28] interesting pod/router-default-5444994796-x2wsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 17:15:35 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Dec 03 17:15:35 crc kubenswrapper[4787]: [+]process-running ok Dec 03 17:15:35 crc kubenswrapper[4787]: healthz check failed Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.237919 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x2wsr" podUID="70235f57-caf7-48d6-ab8e-85230e423cd0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.239606 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" podStartSLOduration=133.239585211 podStartE2EDuration="2m13.239585211s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:35.234710578 +0000 UTC m=+152.052181547" watchObservedRunningTime="2025-12-03 17:15:35.239585211 +0000 UTC m=+152.057056170" Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.251949 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:35 crc kubenswrapper[4787]: E1203 17:15:35.252411 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:35.752395169 +0000 UTC m=+152.569866128 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.257794 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-x7g9b"] Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.296330 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62"] Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.336789 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-f2vjk" podStartSLOduration=133.33676198 podStartE2EDuration="2m13.33676198s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:35.326951634 +0000 UTC m=+152.144422593" watchObservedRunningTime="2025-12-03 17:15:35.33676198 +0000 UTC m=+152.154232939" Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.342391 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms"] Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.342554 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm"] Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.356057 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:35 crc kubenswrapper[4787]: E1203 17:15:35.356280 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:35.856156697 +0000 UTC m=+152.673627656 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.356442 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:35 crc kubenswrapper[4787]: E1203 17:15:35.358584 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:35.858572243 +0000 UTC m=+152.676043202 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.448658 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-rwvrf" podStartSLOduration=134.448631099 podStartE2EDuration="2m14.448631099s" podCreationTimestamp="2025-12-03 17:13:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:35.384400724 +0000 UTC m=+152.201871683" watchObservedRunningTime="2025-12-03 17:15:35.448631099 +0000 UTC m=+152.266102058" Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.458056 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:35 crc kubenswrapper[4787]: E1203 17:15:35.458494 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:35.958478897 +0000 UTC m=+152.775949856 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.514545 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" podStartSLOduration=133.514521559 podStartE2EDuration="2m13.514521559s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:35.49653633 +0000 UTC m=+152.314007289" watchObservedRunningTime="2025-12-03 17:15:35.514521559 +0000 UTC m=+152.331992518" Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.559941 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:35 crc kubenswrapper[4787]: E1203 17:15:35.561516 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:36.061493035 +0000 UTC m=+152.878964014 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.590562 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-mplbm" podStartSLOduration=133.590538114 podStartE2EDuration="2m13.590538114s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:35.590237156 +0000 UTC m=+152.407708115" watchObservedRunningTime="2025-12-03 17:15:35.590538114 +0000 UTC m=+152.408009063" Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.692374 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:35 crc kubenswrapper[4787]: E1203 17:15:35.693686 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:36.193639193 +0000 UTC m=+153.011110152 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.797405 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:35 crc kubenswrapper[4787]: E1203 17:15:35.797993 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:36.297969697 +0000 UTC m=+153.115440656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:35 crc kubenswrapper[4787]: I1203 17:15:35.899177 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:35 crc kubenswrapper[4787]: E1203 17:15:35.899512 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:36.399497215 +0000 UTC m=+153.216968174 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.000538 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:36 crc kubenswrapper[4787]: E1203 17:15:36.000971 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:36.500954131 +0000 UTC m=+153.318425100 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.018621 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" event={"ID":"0466f2b9-1263-4369-80f0-364304f1f2f5","Type":"ContainerStarted","Data":"559d83047522bb45d7edd52fd9dfb3be32a0ee71c7ec5c7b4e2941ad27b874e5"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.021431 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" event={"ID":"97cab125-72b2-454c-9bef-bd70f5eb9654","Type":"ContainerStarted","Data":"bcd0f21e6f4fd258332fba372f7c78cb117561832fa450747dd08be5e5d91d08"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.024063 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-jjvtn" event={"ID":"c09d8632-626c-485d-9e74-65e651546d04","Type":"ContainerStarted","Data":"8523a33a44b41ffd8a0a47fca1cda0632e7075cbf9e8ae93d25052e37da8e628"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.025446 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" event={"ID":"31bdcc6c-aa93-4459-b03f-c559bc726240","Type":"ContainerStarted","Data":"6d30f5b396c116d250e9b34b1f5986680b41fa664f668ad93469cbce6d973fdc"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.037102 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-jjvtn" podStartSLOduration=8.037073972 podStartE2EDuration="8.037073972s" podCreationTimestamp="2025-12-03 17:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:36.035402367 +0000 UTC m=+152.852873326" watchObservedRunningTime="2025-12-03 17:15:36.037073972 +0000 UTC m=+152.854544931" Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.038279 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" event={"ID":"2f958667-8c81-40a0-97c0-0afafae369d2","Type":"ContainerStarted","Data":"0155f0feeddad8869dac52e23bb265226cda01bd8dbf4797e1e5d26b929821e8"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.043095 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" event={"ID":"1af7cc12-ae37-4010-9d9c-28fb02ca03b4","Type":"ContainerStarted","Data":"715b66bdb56e6e2fe553649221d0057f4e09ea16507d34b9ad55b837570d48a5"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.043879 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.045393 4787 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-cj96h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.045503 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" podUID="1af7cc12-ae37-4010-9d9c-28fb02ca03b4" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.048199 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" event={"ID":"e87eccc7-4ea1-4cfe-a219-e8d727fd2051","Type":"ContainerStarted","Data":"a8af2cf02c4e6990d93349ac0737eba8608811a0a349d420d6c7228f63b20801"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.053635 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" event={"ID":"339b0916-61a9-4e97-9fd3-d3a8dc22d23f","Type":"ContainerStarted","Data":"ba97ec45c40501e4ba12063c61ac313f6601cd3be3b06380cb17c24d78367578"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.055716 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp" event={"ID":"440f810f-40c0-4707-afd3-1996592a9cd9","Type":"ContainerStarted","Data":"0340c9cfb9c1ac3052e5dd72062564216253265137ab2da4030cda3ae1ff937c"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.058444 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" podStartSLOduration=134.058432352 podStartE2EDuration="2m14.058432352s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:36.056844969 +0000 UTC m=+152.874315928" watchObservedRunningTime="2025-12-03 17:15:36.058432352 +0000 UTC m=+152.875903311" Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.063206 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cggxn" event={"ID":"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b","Type":"ContainerStarted","Data":"478b21ac51958ba323786db93b489c7fd076a866c057e6e1d3bc41c4bc8dbee4"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.064469 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" event={"ID":"c0c6d238-9876-4f67-b557-94ac4f209332","Type":"ContainerStarted","Data":"c5a585df2184e7c512b37467840c031f477fc9b813d8d947d8944c2b0a39f168"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.064496 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" event={"ID":"c0c6d238-9876-4f67-b557-94ac4f209332","Type":"ContainerStarted","Data":"afeea8163be197345b27780b361b46f36fc93314c8105420e8b865b914edef0d"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.065114 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" event={"ID":"688cf6b7-0d4d-452f-b878-93bd8ff75408","Type":"ContainerStarted","Data":"2a46185600b4b76a7d2157addbd9be9b3856d1f95a89a5d8a32a4c41337e6d25"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.065733 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" event={"ID":"df870037-226f-4286-b128-bbf81e4871b8","Type":"ContainerStarted","Data":"5814bbc36b068bc3bc4b730c04cef6d6796c302a495035d831b865e7902e8515"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.067073 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" event={"ID":"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75","Type":"ContainerStarted","Data":"bd518d37cbfe99df5fa35cdf85b249e1185deb89579c21058d3dbc263fedd71e"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.068651 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" event={"ID":"88a31888-8dbe-45d6-9ce9-0d9c65e9f914","Type":"ContainerStarted","Data":"3cbe9cf515fe23f5df17e899ad700942e6b555ca3b1e392c6798de4a8472af38"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.070401 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" event={"ID":"aba59ac8-ce2a-4321-a2a3-d790af00541c","Type":"ContainerStarted","Data":"14afc223f97a43dc343937ac3b3ac871919c1b7b977c03d3a0b81fa85851b87c"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.073474 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" event={"ID":"552d32f0-c753-4e7d-b0e7-1707e8fa8d26","Type":"ContainerStarted","Data":"4ad26eaa8b5ed2ae3d141f4cea3a3c6d6cefabb54fa713644fd3f8ec0be73b2b"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.075583 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vjhsg" event={"ID":"b3f58901-2824-4993-a0cd-c16e5509894a","Type":"ContainerStarted","Data":"ef3477611477cd7b1e87ac33025cfd9f1349df54d80bda648993e992836ad212"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.082656 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" event={"ID":"684133b3-732f-47c6-b15d-108c3e37397f","Type":"ContainerStarted","Data":"9cbe8ac34d758da238115f28c4c930ec9270479e51b277a9c2d76cba1e2ef25a"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.085135 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" podStartSLOduration=36.085118097 podStartE2EDuration="36.085118097s" podCreationTimestamp="2025-12-03 17:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:36.083195695 +0000 UTC m=+152.900666674" watchObservedRunningTime="2025-12-03 17:15:36.085118097 +0000 UTC m=+152.902589056" Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.089434 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" event={"ID":"27bf1dec-a834-41e8-acdd-e5047002ed0b","Type":"ContainerStarted","Data":"3b09cd16f15e817badb75e4cad2b46d8801239a79fbdea50b530d80849ef445d"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.095229 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6" event={"ID":"ee54b630-23ff-4200-aa84-d3aca72f50e8","Type":"ContainerStarted","Data":"b39ff42d6c50c6b2b303f2e5467ee86424af63d4fcf05bb589622a8e2687876b"} Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.100858 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:36 crc kubenswrapper[4787]: E1203 17:15:36.102089 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:36.602071138 +0000 UTC m=+153.419542087 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.104555 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-fpjsk" Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.141850 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj8b6" podStartSLOduration=134.141836318 podStartE2EDuration="2m14.141836318s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:36.141488308 +0000 UTC m=+152.958959267" watchObservedRunningTime="2025-12-03 17:15:36.141836318 +0000 UTC m=+152.959307277" Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.144060 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-46g9l" podStartSLOduration=134.144051138 podStartE2EDuration="2m14.144051138s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:36.108748589 +0000 UTC m=+152.926219548" watchObservedRunningTime="2025-12-03 17:15:36.144051138 +0000 UTC m=+152.961522097" Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.164601 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-gdftf" podStartSLOduration=134.164565365 podStartE2EDuration="2m14.164565365s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:36.160806603 +0000 UTC m=+152.978277562" watchObservedRunningTime="2025-12-03 17:15:36.164565365 +0000 UTC m=+152.982036324" Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.201854 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:36 crc kubenswrapper[4787]: E1203 17:15:36.204242 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:36.704231823 +0000 UTC m=+153.521702782 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.211498 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" podStartSLOduration=134.211482399 podStartE2EDuration="2m14.211482399s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:36.191102196 +0000 UTC m=+153.008573155" watchObservedRunningTime="2025-12-03 17:15:36.211482399 +0000 UTC m=+153.028953358" Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.223225 4787 patch_prober.go:28] interesting pod/router-default-5444994796-x2wsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 17:15:36 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Dec 03 17:15:36 crc kubenswrapper[4787]: [+]process-running ok Dec 03 17:15:36 crc kubenswrapper[4787]: healthz check failed Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.223315 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x2wsr" podUID="70235f57-caf7-48d6-ab8e-85230e423cd0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.303471 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:36 crc kubenswrapper[4787]: E1203 17:15:36.303843 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:36.803827738 +0000 UTC m=+153.621298697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.406271 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:36 crc kubenswrapper[4787]: E1203 17:15:36.406802 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:36.906786665 +0000 UTC m=+153.724257624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.507991 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:36 crc kubenswrapper[4787]: E1203 17:15:36.508189 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:37.008161848 +0000 UTC m=+153.825632807 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.508336 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:36 crc kubenswrapper[4787]: E1203 17:15:36.508817 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:37.008795575 +0000 UTC m=+153.826266534 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.609798 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:36 crc kubenswrapper[4787]: E1203 17:15:36.610688 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:37.110660563 +0000 UTC m=+153.928131522 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.712258 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:36 crc kubenswrapper[4787]: E1203 17:15:36.712598 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:37.212582861 +0000 UTC m=+154.030053820 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.815078 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:36 crc kubenswrapper[4787]: E1203 17:15:36.815694 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:37.315679301 +0000 UTC m=+154.133150260 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:36 crc kubenswrapper[4787]: I1203 17:15:36.918361 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:36 crc kubenswrapper[4787]: E1203 17:15:36.918796 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:37.418772922 +0000 UTC m=+154.236244061 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.022771 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.023295 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:37.52327567 +0000 UTC m=+154.340746639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.110856 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vjhsg" event={"ID":"b3f58901-2824-4993-a0cd-c16e5509894a","Type":"ContainerStarted","Data":"fb3e207543a006be05af7145b349a61445ac0df11f7673016ecfd4c1d9eaeab6"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.111008 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.112364 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" event={"ID":"0466f2b9-1263-4369-80f0-364304f1f2f5","Type":"ContainerStarted","Data":"7dbf72675505d7b0fb824c12ba0732c0491da80208e6f83484c11b01555b8a9a"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.112886 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.115213 4787 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-h5zms container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" start-of-body= Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.115263 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" podUID="0466f2b9-1263-4369-80f0-364304f1f2f5" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.36:5443/healthz\": dial tcp 10.217.0.36:5443: connect: connection refused" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.117549 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" event={"ID":"339b0916-61a9-4e97-9fd3-d3a8dc22d23f","Type":"ContainerStarted","Data":"afa863f226a2245d3629f4939bd5228cef55c511d30bf3669036e5e9568b2069"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.122979 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" event={"ID":"df870037-226f-4286-b128-bbf81e4871b8","Type":"ContainerStarted","Data":"e60dd573f81cd81391e6c03272b85f5c5732692185e80ffaa64a23269221cf03"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.123969 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.124447 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:37.624431258 +0000 UTC m=+154.441902217 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.136446 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp" event={"ID":"440f810f-40c0-4707-afd3-1996592a9cd9","Type":"ContainerStarted","Data":"dc9ef04739be73731d95dc96a2f794cdc1a0cf0aab7c0b9f9e00672c77eeb61c"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.138919 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" event={"ID":"88a31888-8dbe-45d6-9ce9-0d9c65e9f914","Type":"ContainerStarted","Data":"fc742f0513520f1c3432c33d49844b176ac9f09f37088451671eafc783cd367f"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.139907 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.141135 4787 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-zcnjw container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.141168 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" podUID="88a31888-8dbe-45d6-9ce9-0d9c65e9f914" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.142241 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" event={"ID":"31bdcc6c-aa93-4459-b03f-c559bc726240","Type":"ContainerStarted","Data":"409b573079d2a173d54b1c379fa7b9926dfebe9a39e683a2df52639e9c01882d"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.143947 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" event={"ID":"2f958667-8c81-40a0-97c0-0afafae369d2","Type":"ContainerStarted","Data":"6d59d6201ca0d0d53361fef5b13c8854cf393ec6077565ea44d57c097263aa9b"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.154192 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" event={"ID":"e87eccc7-4ea1-4cfe-a219-e8d727fd2051","Type":"ContainerStarted","Data":"d928d0a030cb702f2823a796270a5628e698daf22603f4ad862875bd8c5d1144"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.156759 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-vjhsg" podStartSLOduration=9.156742106 podStartE2EDuration="9.156742106s" podCreationTimestamp="2025-12-03 17:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:37.154975448 +0000 UTC m=+153.972446407" watchObservedRunningTime="2025-12-03 17:15:37.156742106 +0000 UTC m=+153.974213065" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.163790 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" event={"ID":"688cf6b7-0d4d-452f-b878-93bd8ff75408","Type":"ContainerStarted","Data":"2ad5fc62da488cde537a0610617428bb1d1a3c20a0294a8f87ca4a4a94689f28"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.165649 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" event={"ID":"552d32f0-c753-4e7d-b0e7-1707e8fa8d26","Type":"ContainerStarted","Data":"91a24795bcae14965412c51c0a595a577657f654d756251404cd2950184a2e84"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.167483 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" event={"ID":"97cab125-72b2-454c-9bef-bd70f5eb9654","Type":"ContainerStarted","Data":"3650e664dccc203e4d68724f103437ca6a0feaf23df3e88ea72756a951655ef8"} Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.169659 4787 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-cj96h container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" start-of-body= Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.169711 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" podUID="1af7cc12-ae37-4010-9d9c-28fb02ca03b4" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.18:8443/healthz\": dial tcp 10.217.0.18:8443: connect: connection refused" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.170710 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.172732 4787 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bl7vk container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" start-of-body= Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.172791 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" podUID="684133b3-732f-47c6-b15d-108c3e37397f" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.31:8080/healthz\": dial tcp 10.217.0.31:8080: connect: connection refused" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.180581 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-wb4sr" podStartSLOduration=135.180559883 podStartE2EDuration="2m15.180559883s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:37.177697185 +0000 UTC m=+153.995168144" watchObservedRunningTime="2025-12-03 17:15:37.180559883 +0000 UTC m=+153.998030842" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.202154 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-swngg" podStartSLOduration=135.202121868 podStartE2EDuration="2m15.202121868s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:37.200459753 +0000 UTC m=+154.017930712" watchObservedRunningTime="2025-12-03 17:15:37.202121868 +0000 UTC m=+154.019592827" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.225846 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.227471 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:37.727445436 +0000 UTC m=+154.544916565 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.237373 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" podStartSLOduration=135.237346225 podStartE2EDuration="2m15.237346225s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:37.234513858 +0000 UTC m=+154.051984837" watchObservedRunningTime="2025-12-03 17:15:37.237346225 +0000 UTC m=+154.054817184" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.239757 4787 patch_prober.go:28] interesting pod/router-default-5444994796-x2wsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 17:15:37 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Dec 03 17:15:37 crc kubenswrapper[4787]: [+]process-running ok Dec 03 17:15:37 crc kubenswrapper[4787]: healthz check failed Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.239823 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x2wsr" podUID="70235f57-caf7-48d6-ab8e-85230e423cd0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.259217 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" podStartSLOduration=135.259198609 podStartE2EDuration="2m15.259198609s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:37.258048707 +0000 UTC m=+154.075519676" watchObservedRunningTime="2025-12-03 17:15:37.259198609 +0000 UTC m=+154.076669568" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.279881 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-x7g9b" podStartSLOduration=135.27985821 podStartE2EDuration="2m15.27985821s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:37.276676363 +0000 UTC m=+154.094147322" watchObservedRunningTime="2025-12-03 17:15:37.27985821 +0000 UTC m=+154.097329169" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.296861 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-jgrtv" podStartSLOduration=135.296844141 podStartE2EDuration="2m15.296844141s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:37.294869568 +0000 UTC m=+154.112340527" watchObservedRunningTime="2025-12-03 17:15:37.296844141 +0000 UTC m=+154.114315100" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.325501 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-gg7bd" podStartSLOduration=135.325476249 podStartE2EDuration="2m15.325476249s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:37.323898226 +0000 UTC m=+154.141369185" watchObservedRunningTime="2025-12-03 17:15:37.325476249 +0000 UTC m=+154.142947198" Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.328330 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.328766 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:37.828745048 +0000 UTC m=+154.646216007 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.429563 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.429953 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:37.929931726 +0000 UTC m=+154.747402685 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.530619 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.531056 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.031034683 +0000 UTC m=+154.848505652 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.631573 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.631741 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.131718697 +0000 UTC m=+154.949189666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.631780 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.632170 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.132156499 +0000 UTC m=+154.949627458 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.732534 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.732734 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.23268051 +0000 UTC m=+155.050151479 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.732816 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.733279 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.233268956 +0000 UTC m=+155.050739995 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.835070 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.835227 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.335208155 +0000 UTC m=+155.152679114 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.835358 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.835642 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.335635286 +0000 UTC m=+155.153106245 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.936728 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.936952 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.436921998 +0000 UTC m=+155.254392967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:37 crc kubenswrapper[4787]: I1203 17:15:37.937243 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:37 crc kubenswrapper[4787]: E1203 17:15:37.937652 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.437642247 +0000 UTC m=+155.255113216 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.038914 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:38 crc kubenswrapper[4787]: E1203 17:15:38.039445 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.539428932 +0000 UTC m=+155.356899891 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.141376 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:38 crc kubenswrapper[4787]: E1203 17:15:38.141750 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.641734281 +0000 UTC m=+155.459205240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.192252 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" event={"ID":"df870037-226f-4286-b128-bbf81e4871b8","Type":"ContainerStarted","Data":"ae80a67b7b46f1dd863c033f0de2e759e35be958fb92af6dacbc8cdc484cdf33"} Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.213905 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" event={"ID":"97cab125-72b2-454c-9bef-bd70f5eb9654","Type":"ContainerStarted","Data":"522c82ceaeb00b55384a099616e76e4ffd44e3a854f13db862ed27ef0ce25c6a"} Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.214005 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.224618 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-zk9gm" podStartSLOduration=136.224602532 podStartE2EDuration="2m16.224602532s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:38.222861075 +0000 UTC m=+155.040332044" watchObservedRunningTime="2025-12-03 17:15:38.224602532 +0000 UTC m=+155.042073481" Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.231170 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp" event={"ID":"440f810f-40c0-4707-afd3-1996592a9cd9","Type":"ContainerStarted","Data":"c7543f1393a9e068c508ced935efed6ab9c35504ecf8b87b50f361a1f5d48c23"} Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.240312 4787 patch_prober.go:28] interesting pod/router-default-5444994796-x2wsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 17:15:38 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Dec 03 17:15:38 crc kubenswrapper[4787]: [+]process-running ok Dec 03 17:15:38 crc kubenswrapper[4787]: healthz check failed Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.240386 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x2wsr" podUID="70235f57-caf7-48d6-ab8e-85230e423cd0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.244507 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:38 crc kubenswrapper[4787]: E1203 17:15:38.244845 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.744822671 +0000 UTC m=+155.562293630 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.269766 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cggxn" event={"ID":"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b","Type":"ContainerStarted","Data":"717d78fc0a3ec0263b5f46e3a455bd6ccdc5a77bf348b49c85d95237ca31d197"} Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.271668 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cggxn" event={"ID":"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b","Type":"ContainerStarted","Data":"c15acec63767ca7524a3b038c221aa3a66ae0eb53591cd6aae1131d0ddbe9dbb"} Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.273601 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" podStartSLOduration=136.273587093 podStartE2EDuration="2m16.273587093s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:38.265458162 +0000 UTC m=+155.082929131" watchObservedRunningTime="2025-12-03 17:15:38.273587093 +0000 UTC m=+155.091058052" Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.314555 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sl4cp" podStartSLOduration=136.314528995 podStartE2EDuration="2m16.314528995s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:38.305915181 +0000 UTC m=+155.123386130" watchObservedRunningTime="2025-12-03 17:15:38.314528995 +0000 UTC m=+155.131999954" Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.346866 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:38 crc kubenswrapper[4787]: E1203 17:15:38.347366 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.847346466 +0000 UTC m=+155.664817425 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.355104 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" event={"ID":"c0c6d238-9876-4f67-b557-94ac4f209332","Type":"ContainerStarted","Data":"a53128fac8e686d11f8213f791a915b34f97d356c1d4025adfbfe3354934eac8"} Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.364303 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" event={"ID":"31bdcc6c-aa93-4459-b03f-c559bc726240","Type":"ContainerStarted","Data":"bdfc8d0b0806e1fe105a6b507b1aa4545233f70f8277453efd0e07fb5085bb88"} Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.388664 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.406205 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-zcnjw" Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.414197 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-d72l6" podStartSLOduration=136.414178481 podStartE2EDuration="2m16.414178481s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:38.404949101 +0000 UTC m=+155.222420060" watchObservedRunningTime="2025-12-03 17:15:38.414178481 +0000 UTC m=+155.231649440" Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.449409 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:38 crc kubenswrapper[4787]: E1203 17:15:38.450311 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:38.950295593 +0000 UTC m=+155.767766552 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.524190 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-rb22r" podStartSLOduration=136.524170459 podStartE2EDuration="2m16.524170459s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:38.485952931 +0000 UTC m=+155.303423900" watchObservedRunningTime="2025-12-03 17:15:38.524170459 +0000 UTC m=+155.341641418" Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.553199 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:38 crc kubenswrapper[4787]: E1203 17:15:38.558645 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:39.058629465 +0000 UTC m=+155.876100424 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.650258 4787 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.655111 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:38 crc kubenswrapper[4787]: E1203 17:15:38.655731 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:39.155707822 +0000 UTC m=+155.973178781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.757057 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:38 crc kubenswrapper[4787]: E1203 17:15:38.757507 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:39.257492317 +0000 UTC m=+156.074963276 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.858406 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:38 crc kubenswrapper[4787]: E1203 17:15:38.858984 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-03 17:15:39.358965923 +0000 UTC m=+156.176436882 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.957500 4787 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-03T17:15:38.650291695Z","Handler":null,"Name":""} Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.959594 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:38 crc kubenswrapper[4787]: E1203 17:15:38.959991 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-03 17:15:39.459976697 +0000 UTC m=+156.277447656 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-dw4qj" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.960312 4787 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 03 17:15:38 crc kubenswrapper[4787]: I1203 17:15:38.960346 4787 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.060809 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.090582 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.134954 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-t7zm2"] Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.135817 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.143859 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.161892 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-utilities\") pod \"community-operators-t7zm2\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.161977 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.162163 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-catalog-content\") pod \"community-operators-t7zm2\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.162406 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bbx4z\" (UniqueName: \"kubernetes.io/projected/b4b8d0ee-3c0b-4a5b-9040-b650af634353-kube-api-access-bbx4z\") pod \"community-operators-t7zm2\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.166238 4787 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.166288 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.170071 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7zm2"] Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.225417 4787 patch_prober.go:28] interesting pod/router-default-5444994796-x2wsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 17:15:39 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Dec 03 17:15:39 crc kubenswrapper[4787]: [+]process-running ok Dec 03 17:15:39 crc kubenswrapper[4787]: healthz check failed Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.225470 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x2wsr" podUID="70235f57-caf7-48d6-ab8e-85230e423cd0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.262977 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-utilities\") pod \"community-operators-t7zm2\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.263095 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-catalog-content\") pod \"community-operators-t7zm2\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.263168 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bbx4z\" (UniqueName: \"kubernetes.io/projected/b4b8d0ee-3c0b-4a5b-9040-b650af634353-kube-api-access-bbx4z\") pod \"community-operators-t7zm2\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.264153 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-catalog-content\") pod \"community-operators-t7zm2\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.264314 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-utilities\") pod \"community-operators-t7zm2\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.291664 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bbx4z\" (UniqueName: \"kubernetes.io/projected/b4b8d0ee-3c0b-4a5b-9040-b650af634353-kube-api-access-bbx4z\") pod \"community-operators-t7zm2\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.296196 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-dw4qj\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.321485 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-h5zms" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.334837 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vxhgk"] Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.336031 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.339350 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.351968 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vxhgk"] Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.357479 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.370976 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cggxn" event={"ID":"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b","Type":"ContainerStarted","Data":"0b97ca673a73ac9218459a2f7702ec1ae804c694cc3da64a91d3417aeaea999a"} Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.371042 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-cggxn" event={"ID":"f2d70e9f-74fc-47ca-9e05-4daf81c06c7b","Type":"ContainerStarted","Data":"dc29de5ee4bfd49efa7a6aa2f419cd9531e58658e8199d544ff2fbd2c4acc3dd"} Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.455373 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.485301 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmtr8\" (UniqueName: \"kubernetes.io/projected/ce9df171-a517-4444-9c60-589f83e29e8a-kube-api-access-tmtr8\") pod \"certified-operators-vxhgk\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.485428 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-catalog-content\") pod \"certified-operators-vxhgk\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.485471 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-utilities\") pod \"certified-operators-vxhgk\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.524600 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-cggxn" podStartSLOduration=11.524584242 podStartE2EDuration="11.524584242s" podCreationTimestamp="2025-12-03 17:15:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:39.445435352 +0000 UTC m=+156.262906321" watchObservedRunningTime="2025-12-03 17:15:39.524584242 +0000 UTC m=+156.342055201" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.524950 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hc88l"] Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.528473 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.540180 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hc88l"] Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.553012 4787 patch_prober.go:28] interesting pod/downloads-7954f5f757-xft52 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.553085 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xft52" podUID="ade4b5ab-2584-49e9-9ece-a02a16d79a99" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.553498 4787 patch_prober.go:28] interesting pod/downloads-7954f5f757-xft52 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" start-of-body= Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.553515 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xft52" podUID="ade4b5ab-2584-49e9-9ece-a02a16d79a99" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.12:8080/\": dial tcp 10.217.0.12:8080: connect: connection refused" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.586728 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmtr8\" (UniqueName: \"kubernetes.io/projected/ce9df171-a517-4444-9c60-589f83e29e8a-kube-api-access-tmtr8\") pod \"certified-operators-vxhgk\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.586790 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8skqx\" (UniqueName: \"kubernetes.io/projected/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-kube-api-access-8skqx\") pod \"community-operators-hc88l\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.586815 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-catalog-content\") pod \"certified-operators-vxhgk\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.586844 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-utilities\") pod \"certified-operators-vxhgk\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.586867 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-utilities\") pod \"community-operators-hc88l\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.586907 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-catalog-content\") pod \"community-operators-hc88l\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.587992 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-catalog-content\") pod \"certified-operators-vxhgk\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.588310 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-utilities\") pod \"certified-operators-vxhgk\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.622204 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmtr8\" (UniqueName: \"kubernetes.io/projected/ce9df171-a517-4444-9c60-589f83e29e8a-kube-api-access-tmtr8\") pod \"certified-operators-vxhgk\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.653996 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.676257 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.676769 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.686919 4787 patch_prober.go:28] interesting pod/console-f9d7485db-nm7m8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.686970 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-nm7m8" podUID="a66535aa-a358-4570-ae9f-5dac58121ef2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.687511 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8skqx\" (UniqueName: \"kubernetes.io/projected/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-kube-api-access-8skqx\") pod \"community-operators-hc88l\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.687572 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-utilities\") pod \"community-operators-hc88l\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.687614 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-catalog-content\") pod \"community-operators-hc88l\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.688575 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-catalog-content\") pod \"community-operators-hc88l\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.689006 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-utilities\") pod \"community-operators-hc88l\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.705527 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8skqx\" (UniqueName: \"kubernetes.io/projected/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-kube-api-access-8skqx\") pod \"community-operators-hc88l\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.723820 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6fkg2"] Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.727984 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.734047 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.738504 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-lwzb2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.755440 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6fkg2"] Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.789068 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-catalog-content\") pod \"certified-operators-6fkg2\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.790256 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbddp\" (UniqueName: \"kubernetes.io/projected/4435fc6e-cc89-4399-89dc-c2a87c9b7153-kube-api-access-wbddp\") pod \"certified-operators-6fkg2\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.790438 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-utilities\") pod \"certified-operators-6fkg2\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.793679 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.832504 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.832542 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.859970 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.866383 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.891406 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbddp\" (UniqueName: \"kubernetes.io/projected/4435fc6e-cc89-4399-89dc-c2a87c9b7153-kube-api-access-wbddp\") pod \"certified-operators-6fkg2\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.897259 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-utilities\") pod \"certified-operators-6fkg2\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.897396 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-catalog-content\") pod \"certified-operators-6fkg2\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.898321 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-utilities\") pod \"certified-operators-6fkg2\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.898771 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-catalog-content\") pod \"certified-operators-6fkg2\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.948376 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbddp\" (UniqueName: \"kubernetes.io/projected/4435fc6e-cc89-4399-89dc-c2a87c9b7153-kube-api-access-wbddp\") pod \"certified-operators-6fkg2\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:15:39 crc kubenswrapper[4787]: I1203 17:15:39.959577 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-t7zm2"] Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.017547 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dw4qj"] Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.053662 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vxhgk"] Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.066934 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.221775 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hc88l"] Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.231569 4787 patch_prober.go:28] interesting pod/router-default-5444994796-x2wsr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 03 17:15:40 crc kubenswrapper[4787]: [-]has-synced failed: reason withheld Dec 03 17:15:40 crc kubenswrapper[4787]: [+]process-running ok Dec 03 17:15:40 crc kubenswrapper[4787]: healthz check failed Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.231639 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-x2wsr" podUID="70235f57-caf7-48d6-ab8e-85230e423cd0" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 03 17:15:40 crc kubenswrapper[4787]: W1203 17:15:40.316707 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8c938fa3_8ab3_4d20_b9ef_2fa192313e98.slice/crio-12a5ba350c2090eb34ec9b60f0ab5e91be18f98fc700dbe2ac37ec65b1e9fcaf WatchSource:0}: Error finding container 12a5ba350c2090eb34ec9b60f0ab5e91be18f98fc700dbe2ac37ec65b1e9fcaf: Status 404 returned error can't find the container with id 12a5ba350c2090eb34ec9b60f0ab5e91be18f98fc700dbe2ac37ec65b1e9fcaf Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.353529 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6fkg2"] Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.385978 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc88l" event={"ID":"8c938fa3-8ab3-4d20-b9ef-2fa192313e98","Type":"ContainerStarted","Data":"12a5ba350c2090eb34ec9b60f0ab5e91be18f98fc700dbe2ac37ec65b1e9fcaf"} Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.390848 4787 generic.go:334] "Generic (PLEG): container finished" podID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerID="a7bc0ada5e756e2efe42ba1498340a579e0518f9027865f66ddfc3277875a894" exitCode=0 Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.390955 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7zm2" event={"ID":"b4b8d0ee-3c0b-4a5b-9040-b650af634353","Type":"ContainerDied","Data":"a7bc0ada5e756e2efe42ba1498340a579e0518f9027865f66ddfc3277875a894"} Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.390999 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7zm2" event={"ID":"b4b8d0ee-3c0b-4a5b-9040-b650af634353","Type":"ContainerStarted","Data":"56c7316818e625a5a976cc6c071fff16f56ee2f7953567b527f1d89b597ce4bc"} Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.396933 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.397083 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" event={"ID":"2df4d7cb-f9bb-4ece-9b95-36027a01f017","Type":"ContainerStarted","Data":"7a682029f18a58605dc8172e19f6760b5ac934129a032b613329e7e395623eaa"} Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.397130 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" event={"ID":"2df4d7cb-f9bb-4ece-9b95-36027a01f017","Type":"ContainerStarted","Data":"70b8c57ca75541845296c13338ab4de34f045220318a2938e6e997e16eb6883d"} Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.397368 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.402899 4787 generic.go:334] "Generic (PLEG): container finished" podID="4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75" containerID="bd518d37cbfe99df5fa35cdf85b249e1185deb89579c21058d3dbc263fedd71e" exitCode=0 Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.403044 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" event={"ID":"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75","Type":"ContainerDied","Data":"bd518d37cbfe99df5fa35cdf85b249e1185deb89579c21058d3dbc263fedd71e"} Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.438337 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxhgk" event={"ID":"ce9df171-a517-4444-9c60-589f83e29e8a","Type":"ContainerStarted","Data":"0f880170fca6957d47634a0ab741c099bc0bb2711ef28cdc94ec81bee4ea6cae"} Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.452249 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-wkg8k" Dec 03 17:15:40 crc kubenswrapper[4787]: I1203 17:15:40.463099 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" podStartSLOduration=138.463062494 podStartE2EDuration="2m18.463062494s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:40.452091106 +0000 UTC m=+157.269562075" watchObservedRunningTime="2025-12-03 17:15:40.463062494 +0000 UTC m=+157.280533453" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.125114 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-54jtg"] Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.127063 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.130940 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.137793 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-54jtg"] Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.217838 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.224081 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.224957 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7876\" (UniqueName: \"kubernetes.io/projected/2b45ac06-f7eb-47b8-b1c8-af110a7da844-kube-api-access-w7876\") pod \"redhat-marketplace-54jtg\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.225056 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-utilities\") pod \"redhat-marketplace-54jtg\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.225096 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-catalog-content\") pod \"redhat-marketplace-54jtg\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.325895 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7876\" (UniqueName: \"kubernetes.io/projected/2b45ac06-f7eb-47b8-b1c8-af110a7da844-kube-api-access-w7876\") pod \"redhat-marketplace-54jtg\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.325958 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-utilities\") pod \"redhat-marketplace-54jtg\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.325989 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-catalog-content\") pod \"redhat-marketplace-54jtg\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.327654 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-utilities\") pod \"redhat-marketplace-54jtg\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.327809 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-catalog-content\") pod \"redhat-marketplace-54jtg\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.348155 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7876\" (UniqueName: \"kubernetes.io/projected/2b45ac06-f7eb-47b8-b1c8-af110a7da844-kube-api-access-w7876\") pod \"redhat-marketplace-54jtg\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.388872 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.389793 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.394337 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.394617 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.399345 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.428630 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.428775 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.445282 4787 generic.go:334] "Generic (PLEG): container finished" podID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" containerID="7250b3c1fe8f538d7065c27a327eb5c7b3cba611385f64b0a6ecacf103584d14" exitCode=0 Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.445685 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fkg2" event={"ID":"4435fc6e-cc89-4399-89dc-c2a87c9b7153","Type":"ContainerDied","Data":"7250b3c1fe8f538d7065c27a327eb5c7b3cba611385f64b0a6ecacf103584d14"} Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.445763 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fkg2" event={"ID":"4435fc6e-cc89-4399-89dc-c2a87c9b7153","Type":"ContainerStarted","Data":"09ea0b53bda3fb298c927ad7ecfc44aff7ec4bc5df335c4963475bceb1a99a36"} Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.450430 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.450545 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce9df171-a517-4444-9c60-589f83e29e8a" containerID="2d46071bc3764a213bcfb154ec80446b7313243f16e78da67fb0e18d21c29ab6" exitCode=0 Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.450657 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxhgk" event={"ID":"ce9df171-a517-4444-9c60-589f83e29e8a","Type":"ContainerDied","Data":"2d46071bc3764a213bcfb154ec80446b7313243f16e78da67fb0e18d21c29ab6"} Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.455992 4787 generic.go:334] "Generic (PLEG): container finished" podID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" containerID="57481fa2e2f33f5fa59aeb643105d77d00bb11aaa692f12b1ccd483e8e4e0334" exitCode=0 Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.456262 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc88l" event={"ID":"8c938fa3-8ab3-4d20-b9ef-2fa192313e98","Type":"ContainerDied","Data":"57481fa2e2f33f5fa59aeb643105d77d00bb11aaa692f12b1ccd483e8e4e0334"} Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.466898 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-x2wsr" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.530969 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.531083 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.533766 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.550213 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-llckd"] Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.552559 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.556882 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.574257 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-llckd"] Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.621198 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-cj96h" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.722306 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.746083 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-72cww\" (UniqueName: \"kubernetes.io/projected/ba340e8e-9795-40f8-9690-c224b952b5e6-kube-api-access-72cww\") pod \"redhat-marketplace-llckd\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.746222 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-catalog-content\") pod \"redhat-marketplace-llckd\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.746296 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-utilities\") pod \"redhat-marketplace-llckd\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.844956 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.847382 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-utilities\") pod \"redhat-marketplace-llckd\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.847424 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-72cww\" (UniqueName: \"kubernetes.io/projected/ba340e8e-9795-40f8-9690-c224b952b5e6-kube-api-access-72cww\") pod \"redhat-marketplace-llckd\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.847490 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-catalog-content\") pod \"redhat-marketplace-llckd\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.848207 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-catalog-content\") pod \"redhat-marketplace-llckd\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.848236 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-utilities\") pod \"redhat-marketplace-llckd\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.878622 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-54jtg"] Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.898485 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-72cww\" (UniqueName: \"kubernetes.io/projected/ba340e8e-9795-40f8-9690-c224b952b5e6-kube-api-access-72cww\") pod \"redhat-marketplace-llckd\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.940052 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 17:15:41 crc kubenswrapper[4787]: E1203 17:15:41.940544 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75" containerName="collect-profiles" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.940587 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75" containerName="collect-profiles" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.940757 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75" containerName="collect-profiles" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.941300 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.943880 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.944163 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.945443 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.948936 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tm9xq\" (UniqueName: \"kubernetes.io/projected/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-kube-api-access-tm9xq\") pod \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.949004 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-config-volume\") pod \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.949143 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-secret-volume\") pod \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\" (UID: \"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75\") " Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.950030 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-config-volume" (OuterVolumeSpecName: "config-volume") pod "4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75" (UID: "4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.953908 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-kube-api-access-tm9xq" (OuterVolumeSpecName: "kube-api-access-tm9xq") pod "4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75" (UID: "4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75"). InnerVolumeSpecName "kube-api-access-tm9xq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.968755 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75" (UID: "4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:15:41 crc kubenswrapper[4787]: I1203 17:15:41.970604 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.052548 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a3a6645a-0109-4bd8-b02b-118728e795a9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a3a6645a-0109-4bd8-b02b-118728e795a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.052717 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a3a6645a-0109-4bd8-b02b-118728e795a9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a3a6645a-0109-4bd8-b02b-118728e795a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.052910 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.052974 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tm9xq\" (UniqueName: \"kubernetes.io/projected/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-kube-api-access-tm9xq\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.052989 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.157268 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a3a6645a-0109-4bd8-b02b-118728e795a9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a3a6645a-0109-4bd8-b02b-118728e795a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.157914 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a3a6645a-0109-4bd8-b02b-118728e795a9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a3a6645a-0109-4bd8-b02b-118728e795a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.158501 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a3a6645a-0109-4bd8-b02b-118728e795a9-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"a3a6645a-0109-4bd8-b02b-118728e795a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.255376 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a3a6645a-0109-4bd8-b02b-118728e795a9-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"a3a6645a-0109-4bd8-b02b-118728e795a9\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.285735 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.296310 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 17:15:42 crc kubenswrapper[4787]: W1203 17:15:42.308211 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pode1cbd89c_e5e6_4d9e_bb2c_2717b93161e9.slice/crio-fdbf41a082f66332e2b618d49c7b4e9ea611b6db7686dc608b028d021f959cdd WatchSource:0}: Error finding container fdbf41a082f66332e2b618d49c7b4e9ea611b6db7686dc608b028d021f959cdd: Status 404 returned error can't find the container with id fdbf41a082f66332e2b618d49c7b4e9ea611b6db7686dc608b028d021f959cdd Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.341633 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vnqdl"] Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.343557 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.346432 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.353955 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vnqdl"] Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.463772 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7v49w\" (UniqueName: \"kubernetes.io/projected/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-kube-api-access-7v49w\") pod \"redhat-operators-vnqdl\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.464323 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-utilities\") pod \"redhat-operators-vnqdl\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.464362 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-catalog-content\") pod \"redhat-operators-vnqdl\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.478732 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9","Type":"ContainerStarted","Data":"fdbf41a082f66332e2b618d49c7b4e9ea611b6db7686dc608b028d021f959cdd"} Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.568824 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-utilities\") pod \"redhat-operators-vnqdl\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.568897 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-catalog-content\") pod \"redhat-operators-vnqdl\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.569038 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7v49w\" (UniqueName: \"kubernetes.io/projected/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-kube-api-access-7v49w\") pod \"redhat-operators-vnqdl\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.569716 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-utilities\") pod \"redhat-operators-vnqdl\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.569756 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-catalog-content\") pod \"redhat-operators-vnqdl\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.587757 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7v49w\" (UniqueName: \"kubernetes.io/projected/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-kube-api-access-7v49w\") pod \"redhat-operators-vnqdl\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.602101 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" event={"ID":"4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75","Type":"ContainerDied","Data":"621c87e9f8d3a5756d315d50e416fe9a4032e60dd1b229edac51a1b30032e02c"} Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.602163 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="621c87e9f8d3a5756d315d50e416fe9a4032e60dd1b229edac51a1b30032e02c" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.602253 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.614774 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54jtg" event={"ID":"2b45ac06-f7eb-47b8-b1c8-af110a7da844","Type":"ContainerStarted","Data":"09ba645de41ab67beacfd847a12a921b6497c40200ff16523b1071473494f4be"} Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.698158 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.706974 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.726148 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lg989"] Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.727532 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:15:42 crc kubenswrapper[4787]: W1203 17:15:42.738570 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-poda3a6645a_0109_4bd8_b02b_118728e795a9.slice/crio-850feba52a96787faf58b4890abce3a347b34118974962bc942876c4ed21aa98 WatchSource:0}: Error finding container 850feba52a96787faf58b4890abce3a347b34118974962bc942876c4ed21aa98: Status 404 returned error can't find the container with id 850feba52a96787faf58b4890abce3a347b34118974962bc942876c4ed21aa98 Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.744966 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lg989"] Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.879842 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-utilities\") pod \"redhat-operators-lg989\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.880417 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fnt5f\" (UniqueName: \"kubernetes.io/projected/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-kube-api-access-fnt5f\") pod \"redhat-operators-lg989\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.880539 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-catalog-content\") pod \"redhat-operators-lg989\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.925619 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-llckd"] Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.990725 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-utilities\") pod \"redhat-operators-lg989\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.997620 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fnt5f\" (UniqueName: \"kubernetes.io/projected/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-kube-api-access-fnt5f\") pod \"redhat-operators-lg989\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.998039 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-catalog-content\") pod \"redhat-operators-lg989\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.998748 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-catalog-content\") pod \"redhat-operators-lg989\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:15:42 crc kubenswrapper[4787]: I1203 17:15:42.993964 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-utilities\") pod \"redhat-operators-lg989\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.029506 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fnt5f\" (UniqueName: \"kubernetes.io/projected/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-kube-api-access-fnt5f\") pod \"redhat-operators-lg989\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.106036 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.261680 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vnqdl"] Dec 03 17:15:43 crc kubenswrapper[4787]: W1203 17:15:43.265433 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb25eeef8_0ef5_44b4_acc5_d42ca33b3e43.slice/crio-a9b11c2ff74affbfc48a0c8ed55a78472b53f763233cb6cddd93bfedefa53070 WatchSource:0}: Error finding container a9b11c2ff74affbfc48a0c8ed55a78472b53f763233cb6cddd93bfedefa53070: Status 404 returned error can't find the container with id a9b11c2ff74affbfc48a0c8ed55a78472b53f763233cb6cddd93bfedefa53070 Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.449132 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lg989"] Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.634745 4787 generic.go:334] "Generic (PLEG): container finished" podID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" containerID="74383b00206e80833ae9fb258dc9acce69fa06a1080fb3be2b3d65ed437f2972" exitCode=0 Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.635473 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54jtg" event={"ID":"2b45ac06-f7eb-47b8-b1c8-af110a7da844","Type":"ContainerDied","Data":"74383b00206e80833ae9fb258dc9acce69fa06a1080fb3be2b3d65ed437f2972"} Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.637634 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lg989" event={"ID":"a7b5ebba-5b07-430f-90ea-282fe60f6e5a","Type":"ContainerStarted","Data":"fe65942452f1817fcd636f9133c91a1f2a1b9523a54364c8ba2f33c22ca728c1"} Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.643634 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a3a6645a-0109-4bd8-b02b-118728e795a9","Type":"ContainerStarted","Data":"5bf552ba06f556ee33f1b1ad7d0a62b7bae0890b92350519dac337099d05234e"} Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.643683 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a3a6645a-0109-4bd8-b02b-118728e795a9","Type":"ContainerStarted","Data":"850feba52a96787faf58b4890abce3a347b34118974962bc942876c4ed21aa98"} Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.662838 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llckd" event={"ID":"ba340e8e-9795-40f8-9690-c224b952b5e6","Type":"ContainerStarted","Data":"d004064d11226c05b0c220865c217bd0c2ac9f1993ed5036f62fb9f01a8d9f30"} Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.662898 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llckd" event={"ID":"ba340e8e-9795-40f8-9690-c224b952b5e6","Type":"ContainerStarted","Data":"19e4e6de7a632500619945ea24b0c30b570f325999e629245f968c5e1a7b9e36"} Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.664708 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9","Type":"ContainerStarted","Data":"0e6ae36d68ac1175b0b8aa2117adcafa62d566eff8cc6b0a41bbc6f6eef6655c"} Dec 03 17:15:43 crc kubenswrapper[4787]: I1203 17:15:43.672482 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnqdl" event={"ID":"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43","Type":"ContainerStarted","Data":"a9b11c2ff74affbfc48a0c8ed55a78472b53f763233cb6cddd93bfedefa53070"} Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.424250 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.434897 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/50c2569b-2a14-4112-82e4-afc683aa36a7-metrics-certs\") pod \"network-metrics-daemon-m9tr6\" (UID: \"50c2569b-2a14-4112-82e4-afc683aa36a7\") " pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.684726 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-m9tr6" Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.702884 4787 generic.go:334] "Generic (PLEG): container finished" podID="e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9" containerID="0e6ae36d68ac1175b0b8aa2117adcafa62d566eff8cc6b0a41bbc6f6eef6655c" exitCode=0 Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.702997 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9","Type":"ContainerDied","Data":"0e6ae36d68ac1175b0b8aa2117adcafa62d566eff8cc6b0a41bbc6f6eef6655c"} Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.711512 4787 generic.go:334] "Generic (PLEG): container finished" podID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" containerID="40fe6a8a3887a1007d9f00a5e9d7a3c594b0d59972404137f599289c6d42ba05" exitCode=0 Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.711680 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnqdl" event={"ID":"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43","Type":"ContainerDied","Data":"40fe6a8a3887a1007d9f00a5e9d7a3c594b0d59972404137f599289c6d42ba05"} Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.716061 4787 generic.go:334] "Generic (PLEG): container finished" podID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" containerID="6d62ed84a9e0f2311f39cba1fe0dc3f060487b2c49f1127b33c4ec49310ce6ec" exitCode=0 Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.716170 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lg989" event={"ID":"a7b5ebba-5b07-430f-90ea-282fe60f6e5a","Type":"ContainerDied","Data":"6d62ed84a9e0f2311f39cba1fe0dc3f060487b2c49f1127b33c4ec49310ce6ec"} Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.719506 4787 generic.go:334] "Generic (PLEG): container finished" podID="a3a6645a-0109-4bd8-b02b-118728e795a9" containerID="5bf552ba06f556ee33f1b1ad7d0a62b7bae0890b92350519dac337099d05234e" exitCode=0 Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.719546 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a3a6645a-0109-4bd8-b02b-118728e795a9","Type":"ContainerDied","Data":"5bf552ba06f556ee33f1b1ad7d0a62b7bae0890b92350519dac337099d05234e"} Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.738298 4787 generic.go:334] "Generic (PLEG): container finished" podID="ba340e8e-9795-40f8-9690-c224b952b5e6" containerID="d004064d11226c05b0c220865c217bd0c2ac9f1993ed5036f62fb9f01a8d9f30" exitCode=0 Dec 03 17:15:44 crc kubenswrapper[4787]: I1203 17:15:44.738358 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llckd" event={"ID":"ba340e8e-9795-40f8-9690-c224b952b5e6","Type":"ContainerDied","Data":"d004064d11226c05b0c220865c217bd0c2ac9f1993ed5036f62fb9f01a8d9f30"} Dec 03 17:15:45 crc kubenswrapper[4787]: I1203 17:15:45.020002 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-m9tr6"] Dec 03 17:15:45 crc kubenswrapper[4787]: W1203 17:15:45.033915 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod50c2569b_2a14_4112_82e4_afc683aa36a7.slice/crio-7b51353e84e6a253465ddcb27452d3a223a26955a67d0c24f49b4ae33983592a WatchSource:0}: Error finding container 7b51353e84e6a253465ddcb27452d3a223a26955a67d0c24f49b4ae33983592a: Status 404 returned error can't find the container with id 7b51353e84e6a253465ddcb27452d3a223a26955a67d0c24f49b4ae33983592a Dec 03 17:15:45 crc kubenswrapper[4787]: I1203 17:15:45.745586 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" event={"ID":"50c2569b-2a14-4112-82e4-afc683aa36a7","Type":"ContainerStarted","Data":"7b51353e84e6a253465ddcb27452d3a223a26955a67d0c24f49b4ae33983592a"} Dec 03 17:15:45 crc kubenswrapper[4787]: I1203 17:15:45.962715 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.085415 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.139759 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.145769 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kubelet-dir\") pod \"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9\" (UID: \"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9\") " Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.145886 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9" (UID: "e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.145925 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kube-api-access\") pod \"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9\" (UID: \"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9\") " Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.146237 4787 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.151530 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9" (UID: "e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.246692 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a3a6645a-0109-4bd8-b02b-118728e795a9-kubelet-dir\") pod \"a3a6645a-0109-4bd8-b02b-118728e795a9\" (UID: \"a3a6645a-0109-4bd8-b02b-118728e795a9\") " Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.246749 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a3a6645a-0109-4bd8-b02b-118728e795a9-kube-api-access\") pod \"a3a6645a-0109-4bd8-b02b-118728e795a9\" (UID: \"a3a6645a-0109-4bd8-b02b-118728e795a9\") " Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.246933 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.247393 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a3a6645a-0109-4bd8-b02b-118728e795a9-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "a3a6645a-0109-4bd8-b02b-118728e795a9" (UID: "a3a6645a-0109-4bd8-b02b-118728e795a9"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.261752 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3a6645a-0109-4bd8-b02b-118728e795a9-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "a3a6645a-0109-4bd8-b02b-118728e795a9" (UID: "a3a6645a-0109-4bd8-b02b-118728e795a9"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.349112 4787 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/a3a6645a-0109-4bd8-b02b-118728e795a9-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.349397 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/a3a6645a-0109-4bd8-b02b-118728e795a9-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.710540 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-vjhsg" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.774474 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"a3a6645a-0109-4bd8-b02b-118728e795a9","Type":"ContainerDied","Data":"850feba52a96787faf58b4890abce3a347b34118974962bc942876c4ed21aa98"} Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.774507 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.774526 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="850feba52a96787faf58b4890abce3a347b34118974962bc942876c4ed21aa98" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.776357 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.776375 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9","Type":"ContainerDied","Data":"fdbf41a082f66332e2b618d49c7b4e9ea611b6db7686dc608b028d021f959cdd"} Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.776405 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fdbf41a082f66332e2b618d49c7b4e9ea611b6db7686dc608b028d021f959cdd" Dec 03 17:15:46 crc kubenswrapper[4787]: I1203 17:15:46.779077 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" event={"ID":"50c2569b-2a14-4112-82e4-afc683aa36a7","Type":"ContainerStarted","Data":"897addf879fddcafc1b99e3f05c2f5e0cc5b15d0675c32cce4f806bfdaf035c2"} Dec 03 17:15:47 crc kubenswrapper[4787]: I1203 17:15:47.803448 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-m9tr6" event={"ID":"50c2569b-2a14-4112-82e4-afc683aa36a7","Type":"ContainerStarted","Data":"5e960593a5fd75c12bf068fc2894be0d8dbadc970dc093475d8abff5c224d413"} Dec 03 17:15:47 crc kubenswrapper[4787]: I1203 17:15:47.822351 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-m9tr6" podStartSLOduration=145.822318292 podStartE2EDuration="2m25.822318292s" podCreationTimestamp="2025-12-03 17:13:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:15:47.821817508 +0000 UTC m=+164.639288477" watchObservedRunningTime="2025-12-03 17:15:47.822318292 +0000 UTC m=+164.639789251" Dec 03 17:15:48 crc kubenswrapper[4787]: I1203 17:15:48.990251 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:15:48 crc kubenswrapper[4787]: I1203 17:15:48.990318 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:15:49 crc kubenswrapper[4787]: I1203 17:15:49.557424 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-xft52" Dec 03 17:15:49 crc kubenswrapper[4787]: I1203 17:15:49.677056 4787 patch_prober.go:28] interesting pod/console-f9d7485db-nm7m8 container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" start-of-body= Dec 03 17:15:49 crc kubenswrapper[4787]: I1203 17:15:49.677168 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-nm7m8" podUID="a66535aa-a358-4570-ae9f-5dac58121ef2" containerName="console" probeResult="failure" output="Get \"https://10.217.0.14:8443/health\": dial tcp 10.217.0.14:8443: connect: connection refused" Dec 03 17:15:59 crc kubenswrapper[4787]: I1203 17:15:59.365136 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:15:59 crc kubenswrapper[4787]: I1203 17:15:59.681984 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:15:59 crc kubenswrapper[4787]: I1203 17:15:59.685396 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:16:11 crc kubenswrapper[4787]: I1203 17:16:11.034244 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 03 17:16:11 crc kubenswrapper[4787]: I1203 17:16:11.665689 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" Dec 03 17:16:13 crc kubenswrapper[4787]: E1203 17:16:13.694162 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 17:16:13 crc kubenswrapper[4787]: E1203 17:16:13.694452 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bbx4z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-t7zm2_openshift-marketplace(b4b8d0ee-3c0b-4a5b-9040-b650af634353): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 17:16:13 crc kubenswrapper[4787]: E1203 17:16:13.695689 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-t7zm2" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.451759 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-t7zm2" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.454322 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:67b8d3dc3c0387083a3d461f9de4e6172e465e029123eea5838a423ffeeb1aae: Get \"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:67b8d3dc3c0387083a3d461f9de4e6172e465e029123eea5838a423ffeeb1aae\": context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.454443 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7v49w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-vnqdl_openshift-marketplace(b25eeef8-0ef5-44b4-acc5-d42ca33b3e43): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:67b8d3dc3c0387083a3d461f9de4e6172e465e029123eea5838a423ffeeb1aae: Get \"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:67b8d3dc3c0387083a3d461f9de4e6172e465e029123eea5838a423ffeeb1aae\": context canceled" logger="UnhandledError" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.455640 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: reading blob sha256:67b8d3dc3c0387083a3d461f9de4e6172e465e029123eea5838a423ffeeb1aae: Get \\\"https://registry.redhat.io/v2/redhat/redhat-operator-index/blobs/sha256:67b8d3dc3c0387083a3d461f9de4e6172e465e029123eea5838a423ffeeb1aae\\\": context canceled\"" pod="openshift-marketplace/redhat-operators-vnqdl" podUID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.553572 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.553929 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-8skqx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hc88l_openshift-marketplace(8c938fa3-8ab3-4d20-b9ef-2fa192313e98): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.555074 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hc88l" podUID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.574448 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.574655 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-tmtr8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-vxhgk_openshift-marketplace(ce9df171-a517-4444-9c60-589f83e29e8a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.574853 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.574942 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wbddp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-6fkg2_openshift-marketplace(4435fc6e-cc89-4399-89dc-c2a87c9b7153): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.576046 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-vxhgk" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" Dec 03 17:16:15 crc kubenswrapper[4787]: E1203 17:16:15.576046 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-6fkg2" podUID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" Dec 03 17:16:16 crc kubenswrapper[4787]: E1203 17:16:16.623228 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hc88l" podUID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" Dec 03 17:16:16 crc kubenswrapper[4787]: E1203 17:16:16.623320 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-6fkg2" podUID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" Dec 03 17:16:16 crc kubenswrapper[4787]: E1203 17:16:16.623416 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-vnqdl" podUID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" Dec 03 17:16:16 crc kubenswrapper[4787]: E1203 17:16:16.623228 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-vxhgk" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" Dec 03 17:16:16 crc kubenswrapper[4787]: E1203 17:16:16.637810 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 17:16:16 crc kubenswrapper[4787]: E1203 17:16:16.638009 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-72cww,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-llckd_openshift-marketplace(ba340e8e-9795-40f8-9690-c224b952b5e6): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 17:16:16 crc kubenswrapper[4787]: E1203 17:16:16.639115 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-llckd" podUID="ba340e8e-9795-40f8-9690-c224b952b5e6" Dec 03 17:16:16 crc kubenswrapper[4787]: E1203 17:16:16.647298 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 03 17:16:16 crc kubenswrapper[4787]: E1203 17:16:16.647458 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w7876,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-54jtg_openshift-marketplace(2b45ac06-f7eb-47b8-b1c8-af110a7da844): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 17:16:16 crc kubenswrapper[4787]: E1203 17:16:16.649489 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-54jtg" podUID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.788449 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 17:16:17 crc kubenswrapper[4787]: E1203 17:16:17.788993 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9" containerName="pruner" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.789007 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9" containerName="pruner" Dec 03 17:16:17 crc kubenswrapper[4787]: E1203 17:16:17.789044 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3a6645a-0109-4bd8-b02b-118728e795a9" containerName="pruner" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.789053 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3a6645a-0109-4bd8-b02b-118728e795a9" containerName="pruner" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.789184 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1cbd89c-e5e6-4d9e-bb2c-2717b93161e9" containerName="pruner" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.789200 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3a6645a-0109-4bd8-b02b-118728e795a9" containerName="pruner" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.789642 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.791831 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.792858 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.793163 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.869131 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.869231 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.970205 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.970308 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 17:16:17 crc kubenswrapper[4787]: I1203 17:16:17.970417 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 17:16:18 crc kubenswrapper[4787]: I1203 17:16:18.002970 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 17:16:18 crc kubenswrapper[4787]: I1203 17:16:18.106056 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 17:16:18 crc kubenswrapper[4787]: I1203 17:16:18.989575 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:16:18 crc kubenswrapper[4787]: I1203 17:16:18.989896 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:16:19 crc kubenswrapper[4787]: E1203 17:16:19.989394 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-54jtg" podUID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" Dec 03 17:16:19 crc kubenswrapper[4787]: E1203 17:16:19.989470 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-llckd" podUID="ba340e8e-9795-40f8-9690-c224b952b5e6" Dec 03 17:16:20 crc kubenswrapper[4787]: E1203 17:16:20.011808 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 03 17:16:20 crc kubenswrapper[4787]: E1203 17:16:20.012013 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fnt5f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-lg989_openshift-marketplace(a7b5ebba-5b07-430f-90ea-282fe60f6e5a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 03 17:16:20 crc kubenswrapper[4787]: E1203 17:16:20.013505 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-lg989" podUID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" Dec 03 17:16:20 crc kubenswrapper[4787]: E1203 17:16:20.104327 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-lg989" podUID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" Dec 03 17:16:20 crc kubenswrapper[4787]: I1203 17:16:20.426287 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 03 17:16:21 crc kubenswrapper[4787]: I1203 17:16:21.109306 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47","Type":"ContainerStarted","Data":"05737199c5431067652084bb169a26f788a97149ab06e3bc61cfff0794a25113"} Dec 03 17:16:21 crc kubenswrapper[4787]: I1203 17:16:21.109623 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47","Type":"ContainerStarted","Data":"a55721ff95e7e2574550930cd292493c0052fbc580afa68f9d7b67cd5e366ea6"} Dec 03 17:16:21 crc kubenswrapper[4787]: I1203 17:16:21.128066 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=4.128046054 podStartE2EDuration="4.128046054s" podCreationTimestamp="2025-12-03 17:16:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:16:21.124058098 +0000 UTC m=+197.941529077" watchObservedRunningTime="2025-12-03 17:16:21.128046054 +0000 UTC m=+197.945517023" Dec 03 17:16:22 crc kubenswrapper[4787]: I1203 17:16:22.116009 4787 generic.go:334] "Generic (PLEG): container finished" podID="7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47" containerID="05737199c5431067652084bb169a26f788a97149ab06e3bc61cfff0794a25113" exitCode=0 Dec 03 17:16:22 crc kubenswrapper[4787]: I1203 17:16:22.116216 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47","Type":"ContainerDied","Data":"05737199c5431067652084bb169a26f788a97149ab06e3bc61cfff0794a25113"} Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.356459 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.540457 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kubelet-dir\") pod \"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47\" (UID: \"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47\") " Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.540587 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kube-api-access\") pod \"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47\" (UID: \"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47\") " Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.541873 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47" (UID: "7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.547366 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47" (UID: "7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.574058 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 17:16:23 crc kubenswrapper[4787]: E1203 17:16:23.574251 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47" containerName="pruner" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.574262 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47" containerName="pruner" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.574355 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47" containerName="pruner" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.574701 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.587859 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.641867 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.641929 4787 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.743341 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kube-api-access\") pod \"installer-9-crc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.743406 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.743457 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-var-lock\") pod \"installer-9-crc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.845354 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kube-api-access\") pod \"installer-9-crc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.845433 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.845499 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-var-lock\") pod \"installer-9-crc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.845657 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-var-lock\") pod \"installer-9-crc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.845718 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kubelet-dir\") pod \"installer-9-crc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.864449 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kube-api-access\") pod \"installer-9-crc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:16:23 crc kubenswrapper[4787]: I1203 17:16:23.900058 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:16:24 crc kubenswrapper[4787]: I1203 17:16:24.076717 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 03 17:16:24 crc kubenswrapper[4787]: W1203 17:16:24.087534 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod2cfbeef7_cc82_419e_ae28_c6eef3202ebc.slice/crio-262421ddf72cca4d12499646acb1eca141f3117d75bef12aa683ff1df2251f00 WatchSource:0}: Error finding container 262421ddf72cca4d12499646acb1eca141f3117d75bef12aa683ff1df2251f00: Status 404 returned error can't find the container with id 262421ddf72cca4d12499646acb1eca141f3117d75bef12aa683ff1df2251f00 Dec 03 17:16:24 crc kubenswrapper[4787]: I1203 17:16:24.126051 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2cfbeef7-cc82-419e-ae28-c6eef3202ebc","Type":"ContainerStarted","Data":"262421ddf72cca4d12499646acb1eca141f3117d75bef12aa683ff1df2251f00"} Dec 03 17:16:24 crc kubenswrapper[4787]: I1203 17:16:24.127538 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7537bc9b-51b3-4c5f-aafe-5bfe8c21eb47","Type":"ContainerDied","Data":"a55721ff95e7e2574550930cd292493c0052fbc580afa68f9d7b67cd5e366ea6"} Dec 03 17:16:24 crc kubenswrapper[4787]: I1203 17:16:24.127561 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a55721ff95e7e2574550930cd292493c0052fbc580afa68f9d7b67cd5e366ea6" Dec 03 17:16:24 crc kubenswrapper[4787]: I1203 17:16:24.127629 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 03 17:16:25 crc kubenswrapper[4787]: I1203 17:16:25.133770 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2cfbeef7-cc82-419e-ae28-c6eef3202ebc","Type":"ContainerStarted","Data":"615ce9f7de0380159065e5ecc8d1f573f0d7a38d58733fc2e437263d87a91ae2"} Dec 03 17:16:25 crc kubenswrapper[4787]: I1203 17:16:25.149910 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.149891278 podStartE2EDuration="2.149891278s" podCreationTimestamp="2025-12-03 17:16:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:16:25.147341398 +0000 UTC m=+201.964812367" watchObservedRunningTime="2025-12-03 17:16:25.149891278 +0000 UTC m=+201.967362237" Dec 03 17:16:28 crc kubenswrapper[4787]: I1203 17:16:28.150666 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7zm2" event={"ID":"b4b8d0ee-3c0b-4a5b-9040-b650af634353","Type":"ContainerStarted","Data":"5a5e127cd3ad536e0809114103e75e6b03b7bbb4ecdb35da2365d4cff2dd2849"} Dec 03 17:16:29 crc kubenswrapper[4787]: I1203 17:16:29.156598 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7zm2" event={"ID":"b4b8d0ee-3c0b-4a5b-9040-b650af634353","Type":"ContainerDied","Data":"5a5e127cd3ad536e0809114103e75e6b03b7bbb4ecdb35da2365d4cff2dd2849"} Dec 03 17:16:29 crc kubenswrapper[4787]: I1203 17:16:29.156723 4787 generic.go:334] "Generic (PLEG): container finished" podID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerID="5a5e127cd3ad536e0809114103e75e6b03b7bbb4ecdb35da2365d4cff2dd2849" exitCode=0 Dec 03 17:16:30 crc kubenswrapper[4787]: I1203 17:16:30.166967 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7zm2" event={"ID":"b4b8d0ee-3c0b-4a5b-9040-b650af634353","Type":"ContainerStarted","Data":"e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b"} Dec 03 17:16:30 crc kubenswrapper[4787]: I1203 17:16:30.184007 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-t7zm2" podStartSLOduration=1.912485078 podStartE2EDuration="51.183988754s" podCreationTimestamp="2025-12-03 17:15:39 +0000 UTC" firstStartedPulling="2025-12-03 17:15:40.39663105 +0000 UTC m=+157.214102009" lastFinishedPulling="2025-12-03 17:16:29.668134726 +0000 UTC m=+206.485605685" observedRunningTime="2025-12-03 17:16:30.18070643 +0000 UTC m=+206.998177389" watchObservedRunningTime="2025-12-03 17:16:30.183988754 +0000 UTC m=+207.001459713" Dec 03 17:16:31 crc kubenswrapper[4787]: I1203 17:16:31.176427 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnqdl" event={"ID":"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43","Type":"ContainerStarted","Data":"e0478542f27129f9f90f5aea14314e99e262284d82f8da7aa406f3977b7e8278"} Dec 03 17:16:32 crc kubenswrapper[4787]: I1203 17:16:32.183453 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxhgk" event={"ID":"ce9df171-a517-4444-9c60-589f83e29e8a","Type":"ContainerStarted","Data":"0108690ad93f843c0a2d787ae052bc7127ddfe527e17b6301cf693aa64dad5dc"} Dec 03 17:16:32 crc kubenswrapper[4787]: I1203 17:16:32.185638 4787 generic.go:334] "Generic (PLEG): container finished" podID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" containerID="e0478542f27129f9f90f5aea14314e99e262284d82f8da7aa406f3977b7e8278" exitCode=0 Dec 03 17:16:32 crc kubenswrapper[4787]: I1203 17:16:32.185684 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnqdl" event={"ID":"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43","Type":"ContainerDied","Data":"e0478542f27129f9f90f5aea14314e99e262284d82f8da7aa406f3977b7e8278"} Dec 03 17:16:33 crc kubenswrapper[4787]: I1203 17:16:33.197470 4787 generic.go:334] "Generic (PLEG): container finished" podID="ba340e8e-9795-40f8-9690-c224b952b5e6" containerID="513d315f6926494d40aad1a45794fe2296ba33dd3cce90660437f93fa754f75a" exitCode=0 Dec 03 17:16:33 crc kubenswrapper[4787]: I1203 17:16:33.197536 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llckd" event={"ID":"ba340e8e-9795-40f8-9690-c224b952b5e6","Type":"ContainerDied","Data":"513d315f6926494d40aad1a45794fe2296ba33dd3cce90660437f93fa754f75a"} Dec 03 17:16:33 crc kubenswrapper[4787]: I1203 17:16:33.201071 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fkg2" event={"ID":"4435fc6e-cc89-4399-89dc-c2a87c9b7153","Type":"ContainerStarted","Data":"2ec51a2b80012bc2e8b1b803b353674207b87b0d4ad49715d1b4f79b41224623"} Dec 03 17:16:33 crc kubenswrapper[4787]: I1203 17:16:33.203568 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce9df171-a517-4444-9c60-589f83e29e8a" containerID="0108690ad93f843c0a2d787ae052bc7127ddfe527e17b6301cf693aa64dad5dc" exitCode=0 Dec 03 17:16:33 crc kubenswrapper[4787]: I1203 17:16:33.203655 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxhgk" event={"ID":"ce9df171-a517-4444-9c60-589f83e29e8a","Type":"ContainerDied","Data":"0108690ad93f843c0a2d787ae052bc7127ddfe527e17b6301cf693aa64dad5dc"} Dec 03 17:16:33 crc kubenswrapper[4787]: I1203 17:16:33.207509 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnqdl" event={"ID":"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43","Type":"ContainerStarted","Data":"4381fe49861a72228d6c3e8f60f3fbd18b8c01d4903b8becb40228b92d50a150"} Dec 03 17:16:33 crc kubenswrapper[4787]: I1203 17:16:33.209840 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc88l" event={"ID":"8c938fa3-8ab3-4d20-b9ef-2fa192313e98","Type":"ContainerStarted","Data":"193dbbb09ecba3dd05c9a03a2a01e7a3499d3b86533c1c4e8367caf4fa8e6e56"} Dec 03 17:16:33 crc kubenswrapper[4787]: I1203 17:16:33.263545 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vnqdl" podStartSLOduration=3.374906414 podStartE2EDuration="51.26352962s" podCreationTimestamp="2025-12-03 17:15:42 +0000 UTC" firstStartedPulling="2025-12-03 17:15:44.713240211 +0000 UTC m=+161.530711170" lastFinishedPulling="2025-12-03 17:16:32.601863417 +0000 UTC m=+209.419334376" observedRunningTime="2025-12-03 17:16:33.262215188 +0000 UTC m=+210.079686147" watchObservedRunningTime="2025-12-03 17:16:33.26352962 +0000 UTC m=+210.081000579" Dec 03 17:16:34 crc kubenswrapper[4787]: I1203 17:16:34.217427 4787 generic.go:334] "Generic (PLEG): container finished" podID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" containerID="2ec51a2b80012bc2e8b1b803b353674207b87b0d4ad49715d1b4f79b41224623" exitCode=0 Dec 03 17:16:34 crc kubenswrapper[4787]: I1203 17:16:34.217486 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fkg2" event={"ID":"4435fc6e-cc89-4399-89dc-c2a87c9b7153","Type":"ContainerDied","Data":"2ec51a2b80012bc2e8b1b803b353674207b87b0d4ad49715d1b4f79b41224623"} Dec 03 17:16:34 crc kubenswrapper[4787]: I1203 17:16:34.220355 4787 generic.go:334] "Generic (PLEG): container finished" podID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" containerID="193dbbb09ecba3dd05c9a03a2a01e7a3499d3b86533c1c4e8367caf4fa8e6e56" exitCode=0 Dec 03 17:16:34 crc kubenswrapper[4787]: I1203 17:16:34.220383 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc88l" event={"ID":"8c938fa3-8ab3-4d20-b9ef-2fa192313e98","Type":"ContainerDied","Data":"193dbbb09ecba3dd05c9a03a2a01e7a3499d3b86533c1c4e8367caf4fa8e6e56"} Dec 03 17:16:35 crc kubenswrapper[4787]: I1203 17:16:35.240814 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fkg2" event={"ID":"4435fc6e-cc89-4399-89dc-c2a87c9b7153","Type":"ContainerStarted","Data":"42d0064fc4f76a2c8d581eb2f7b9f11a288644718f5446f889a5866c6aae2c18"} Dec 03 17:16:35 crc kubenswrapper[4787]: I1203 17:16:35.244140 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxhgk" event={"ID":"ce9df171-a517-4444-9c60-589f83e29e8a","Type":"ContainerStarted","Data":"31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30"} Dec 03 17:16:35 crc kubenswrapper[4787]: I1203 17:16:35.246782 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc88l" event={"ID":"8c938fa3-8ab3-4d20-b9ef-2fa192313e98","Type":"ContainerStarted","Data":"16f8f7b3c41a08f81c78dbc873f9713ee9864e61abff7bf7e2650c3b45442477"} Dec 03 17:16:35 crc kubenswrapper[4787]: I1203 17:16:35.248981 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lg989" event={"ID":"a7b5ebba-5b07-430f-90ea-282fe60f6e5a","Type":"ContainerStarted","Data":"5c29234e8ce390ad2ae3dfc60ffb3110c233d0c8228e8d07875f273b8927d6c7"} Dec 03 17:16:35 crc kubenswrapper[4787]: I1203 17:16:35.250699 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llckd" event={"ID":"ba340e8e-9795-40f8-9690-c224b952b5e6","Type":"ContainerStarted","Data":"066f0274138220a0a7fae618b82a115d07c86427f738ed458e4506000995a7b1"} Dec 03 17:16:35 crc kubenswrapper[4787]: I1203 17:16:35.263235 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6fkg2" podStartSLOduration=3.010075529 podStartE2EDuration="56.263215253s" podCreationTimestamp="2025-12-03 17:15:39 +0000 UTC" firstStartedPulling="2025-12-03 17:15:41.448737288 +0000 UTC m=+158.266208247" lastFinishedPulling="2025-12-03 17:16:34.701877012 +0000 UTC m=+211.519347971" observedRunningTime="2025-12-03 17:16:35.262342666 +0000 UTC m=+212.079813625" watchObservedRunningTime="2025-12-03 17:16:35.263215253 +0000 UTC m=+212.080686212" Dec 03 17:16:35 crc kubenswrapper[4787]: I1203 17:16:35.295111 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hc88l" podStartSLOduration=3.122559474 podStartE2EDuration="56.295096698s" podCreationTimestamp="2025-12-03 17:15:39 +0000 UTC" firstStartedPulling="2025-12-03 17:15:41.464924748 +0000 UTC m=+158.282395707" lastFinishedPulling="2025-12-03 17:16:34.637461972 +0000 UTC m=+211.454932931" observedRunningTime="2025-12-03 17:16:35.292762205 +0000 UTC m=+212.110233184" watchObservedRunningTime="2025-12-03 17:16:35.295096698 +0000 UTC m=+212.112567657" Dec 03 17:16:35 crc kubenswrapper[4787]: I1203 17:16:35.313038 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-llckd" podStartSLOduration=4.769996073 podStartE2EDuration="54.313011483s" podCreationTimestamp="2025-12-03 17:15:41 +0000 UTC" firstStartedPulling="2025-12-03 17:15:44.741522689 +0000 UTC m=+161.558993648" lastFinishedPulling="2025-12-03 17:16:34.284538099 +0000 UTC m=+211.102009058" observedRunningTime="2025-12-03 17:16:35.308902573 +0000 UTC m=+212.126373532" watchObservedRunningTime="2025-12-03 17:16:35.313011483 +0000 UTC m=+212.130482442" Dec 03 17:16:35 crc kubenswrapper[4787]: I1203 17:16:35.353416 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vxhgk" podStartSLOduration=3.60299312 podStartE2EDuration="56.353396116s" podCreationTimestamp="2025-12-03 17:15:39 +0000 UTC" firstStartedPulling="2025-12-03 17:15:41.452625304 +0000 UTC m=+158.270096273" lastFinishedPulling="2025-12-03 17:16:34.20302831 +0000 UTC m=+211.020499269" observedRunningTime="2025-12-03 17:16:35.351874818 +0000 UTC m=+212.169345787" watchObservedRunningTime="2025-12-03 17:16:35.353396116 +0000 UTC m=+212.170867075" Dec 03 17:16:36 crc kubenswrapper[4787]: I1203 17:16:36.279283 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54jtg" event={"ID":"2b45ac06-f7eb-47b8-b1c8-af110a7da844","Type":"ContainerStarted","Data":"507d5d030696c7bf38ef96a952b0b59248fffe7ae3780565c489d2a12e8cfc9d"} Dec 03 17:16:37 crc kubenswrapper[4787]: I1203 17:16:37.286560 4787 generic.go:334] "Generic (PLEG): container finished" podID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" containerID="507d5d030696c7bf38ef96a952b0b59248fffe7ae3780565c489d2a12e8cfc9d" exitCode=0 Dec 03 17:16:37 crc kubenswrapper[4787]: I1203 17:16:37.286647 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54jtg" event={"ID":"2b45ac06-f7eb-47b8-b1c8-af110a7da844","Type":"ContainerDied","Data":"507d5d030696c7bf38ef96a952b0b59248fffe7ae3780565c489d2a12e8cfc9d"} Dec 03 17:16:37 crc kubenswrapper[4787]: I1203 17:16:37.289778 4787 generic.go:334] "Generic (PLEG): container finished" podID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" containerID="5c29234e8ce390ad2ae3dfc60ffb3110c233d0c8228e8d07875f273b8927d6c7" exitCode=0 Dec 03 17:16:37 crc kubenswrapper[4787]: I1203 17:16:37.289814 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lg989" event={"ID":"a7b5ebba-5b07-430f-90ea-282fe60f6e5a","Type":"ContainerDied","Data":"5c29234e8ce390ad2ae3dfc60ffb3110c233d0c8228e8d07875f273b8927d6c7"} Dec 03 17:16:39 crc kubenswrapper[4787]: I1203 17:16:39.457605 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:16:39 crc kubenswrapper[4787]: I1203 17:16:39.457657 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:16:39 crc kubenswrapper[4787]: I1203 17:16:39.655164 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:16:39 crc kubenswrapper[4787]: I1203 17:16:39.655240 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:16:39 crc kubenswrapper[4787]: I1203 17:16:39.866997 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:16:39 crc kubenswrapper[4787]: I1203 17:16:39.867310 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:16:40 crc kubenswrapper[4787]: I1203 17:16:40.037182 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:16:40 crc kubenswrapper[4787]: I1203 17:16:40.048550 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:16:40 crc kubenswrapper[4787]: I1203 17:16:40.050686 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:16:40 crc kubenswrapper[4787]: I1203 17:16:40.068095 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:16:40 crc kubenswrapper[4787]: I1203 17:16:40.068154 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:16:40 crc kubenswrapper[4787]: I1203 17:16:40.114539 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:16:40 crc kubenswrapper[4787]: I1203 17:16:40.343803 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:16:40 crc kubenswrapper[4787]: I1203 17:16:40.353956 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:16:40 crc kubenswrapper[4787]: I1203 17:16:40.356997 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:16:40 crc kubenswrapper[4787]: I1203 17:16:40.372604 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:16:41 crc kubenswrapper[4787]: I1203 17:16:41.971166 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:16:41 crc kubenswrapper[4787]: I1203 17:16:41.971278 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:16:42 crc kubenswrapper[4787]: I1203 17:16:42.043706 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:16:42 crc kubenswrapper[4787]: I1203 17:16:42.383531 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:16:42 crc kubenswrapper[4787]: I1203 17:16:42.699584 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:16:42 crc kubenswrapper[4787]: I1203 17:16:42.699667 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:16:42 crc kubenswrapper[4787]: I1203 17:16:42.767422 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:16:42 crc kubenswrapper[4787]: I1203 17:16:42.785802 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6fkg2"] Dec 03 17:16:42 crc kubenswrapper[4787]: I1203 17:16:42.786074 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6fkg2" podUID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" containerName="registry-server" containerID="cri-o://42d0064fc4f76a2c8d581eb2f7b9f11a288644718f5446f889a5866c6aae2c18" gracePeriod=2 Dec 03 17:16:43 crc kubenswrapper[4787]: I1203 17:16:43.370767 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:16:44 crc kubenswrapper[4787]: I1203 17:16:44.189960 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hc88l"] Dec 03 17:16:44 crc kubenswrapper[4787]: I1203 17:16:44.190304 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hc88l" podUID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" containerName="registry-server" containerID="cri-o://16f8f7b3c41a08f81c78dbc873f9713ee9864e61abff7bf7e2650c3b45442477" gracePeriod=2 Dec 03 17:16:45 crc kubenswrapper[4787]: I1203 17:16:45.190500 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-llckd"] Dec 03 17:16:45 crc kubenswrapper[4787]: I1203 17:16:45.191223 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-llckd" podUID="ba340e8e-9795-40f8-9690-c224b952b5e6" containerName="registry-server" containerID="cri-o://066f0274138220a0a7fae618b82a115d07c86427f738ed458e4506000995a7b1" gracePeriod=2 Dec 03 17:16:45 crc kubenswrapper[4787]: I1203 17:16:45.345486 4787 generic.go:334] "Generic (PLEG): container finished" podID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" containerID="42d0064fc4f76a2c8d581eb2f7b9f11a288644718f5446f889a5866c6aae2c18" exitCode=0 Dec 03 17:16:45 crc kubenswrapper[4787]: I1203 17:16:45.345529 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fkg2" event={"ID":"4435fc6e-cc89-4399-89dc-c2a87c9b7153","Type":"ContainerDied","Data":"42d0064fc4f76a2c8d581eb2f7b9f11a288644718f5446f889a5866c6aae2c18"} Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.360264 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lg989" event={"ID":"a7b5ebba-5b07-430f-90ea-282fe60f6e5a","Type":"ContainerStarted","Data":"58eef3295e8300e4ba409fd5dab42843184602c96e3536f3042d95e8dc9295e6"} Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.362901 4787 generic.go:334] "Generic (PLEG): container finished" podID="ba340e8e-9795-40f8-9690-c224b952b5e6" containerID="066f0274138220a0a7fae618b82a115d07c86427f738ed458e4506000995a7b1" exitCode=0 Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.362974 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llckd" event={"ID":"ba340e8e-9795-40f8-9690-c224b952b5e6","Type":"ContainerDied","Data":"066f0274138220a0a7fae618b82a115d07c86427f738ed458e4506000995a7b1"} Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.364722 4787 generic.go:334] "Generic (PLEG): container finished" podID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" containerID="16f8f7b3c41a08f81c78dbc873f9713ee9864e61abff7bf7e2650c3b45442477" exitCode=0 Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.364747 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc88l" event={"ID":"8c938fa3-8ab3-4d20-b9ef-2fa192313e98","Type":"ContainerDied","Data":"16f8f7b3c41a08f81c78dbc873f9713ee9864e61abff7bf7e2650c3b45442477"} Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.576367 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.722614 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-catalog-content\") pod \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.722680 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-utilities\") pod \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.722799 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbddp\" (UniqueName: \"kubernetes.io/projected/4435fc6e-cc89-4399-89dc-c2a87c9b7153-kube-api-access-wbddp\") pod \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\" (UID: \"4435fc6e-cc89-4399-89dc-c2a87c9b7153\") " Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.724233 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-utilities" (OuterVolumeSpecName: "utilities") pod "4435fc6e-cc89-4399-89dc-c2a87c9b7153" (UID: "4435fc6e-cc89-4399-89dc-c2a87c9b7153"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.729910 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4435fc6e-cc89-4399-89dc-c2a87c9b7153-kube-api-access-wbddp" (OuterVolumeSpecName: "kube-api-access-wbddp") pod "4435fc6e-cc89-4399-89dc-c2a87c9b7153" (UID: "4435fc6e-cc89-4399-89dc-c2a87c9b7153"). InnerVolumeSpecName "kube-api-access-wbddp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.773958 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4435fc6e-cc89-4399-89dc-c2a87c9b7153" (UID: "4435fc6e-cc89-4399-89dc-c2a87c9b7153"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.823715 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.823748 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4435fc6e-cc89-4399-89dc-c2a87c9b7153-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:47 crc kubenswrapper[4787]: I1203 17:16:47.823757 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbddp\" (UniqueName: \"kubernetes.io/projected/4435fc6e-cc89-4399-89dc-c2a87c9b7153-kube-api-access-wbddp\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.373341 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6fkg2" event={"ID":"4435fc6e-cc89-4399-89dc-c2a87c9b7153","Type":"ContainerDied","Data":"09ea0b53bda3fb298c927ad7ecfc44aff7ec4bc5df335c4963475bceb1a99a36"} Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.373424 4787 scope.go:117] "RemoveContainer" containerID="42d0064fc4f76a2c8d581eb2f7b9f11a288644718f5446f889a5866c6aae2c18" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.373361 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6fkg2" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.395450 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lg989" podStartSLOduration=6.632308133 podStartE2EDuration="1m6.395429673s" podCreationTimestamp="2025-12-03 17:15:42 +0000 UTC" firstStartedPulling="2025-12-03 17:15:44.718247467 +0000 UTC m=+161.535718426" lastFinishedPulling="2025-12-03 17:16:44.481369007 +0000 UTC m=+221.298839966" observedRunningTime="2025-12-03 17:16:48.39406412 +0000 UTC m=+225.211535129" watchObservedRunningTime="2025-12-03 17:16:48.395429673 +0000 UTC m=+225.212900632" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.406876 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6fkg2"] Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.418048 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6fkg2"] Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.496345 4787 scope.go:117] "RemoveContainer" containerID="2ec51a2b80012bc2e8b1b803b353674207b87b0d4ad49715d1b4f79b41224623" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.548637 4787 scope.go:117] "RemoveContainer" containerID="7250b3c1fe8f538d7065c27a327eb5c7b3cba611385f64b0a6ecacf103584d14" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.770932 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.774263 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.939473 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-utilities\") pod \"ba340e8e-9795-40f8-9690-c224b952b5e6\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.939775 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8skqx\" (UniqueName: \"kubernetes.io/projected/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-kube-api-access-8skqx\") pod \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.939861 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-utilities\") pod \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.939903 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-72cww\" (UniqueName: \"kubernetes.io/projected/ba340e8e-9795-40f8-9690-c224b952b5e6-kube-api-access-72cww\") pod \"ba340e8e-9795-40f8-9690-c224b952b5e6\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.939939 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-catalog-content\") pod \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\" (UID: \"8c938fa3-8ab3-4d20-b9ef-2fa192313e98\") " Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.939973 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-catalog-content\") pod \"ba340e8e-9795-40f8-9690-c224b952b5e6\" (UID: \"ba340e8e-9795-40f8-9690-c224b952b5e6\") " Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.941730 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-utilities" (OuterVolumeSpecName: "utilities") pod "8c938fa3-8ab3-4d20-b9ef-2fa192313e98" (UID: "8c938fa3-8ab3-4d20-b9ef-2fa192313e98"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.942006 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-utilities" (OuterVolumeSpecName: "utilities") pod "ba340e8e-9795-40f8-9690-c224b952b5e6" (UID: "ba340e8e-9795-40f8-9690-c224b952b5e6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.945573 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba340e8e-9795-40f8-9690-c224b952b5e6-kube-api-access-72cww" (OuterVolumeSpecName: "kube-api-access-72cww") pod "ba340e8e-9795-40f8-9690-c224b952b5e6" (UID: "ba340e8e-9795-40f8-9690-c224b952b5e6"). InnerVolumeSpecName "kube-api-access-72cww". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.951177 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-kube-api-access-8skqx" (OuterVolumeSpecName: "kube-api-access-8skqx") pod "8c938fa3-8ab3-4d20-b9ef-2fa192313e98" (UID: "8c938fa3-8ab3-4d20-b9ef-2fa192313e98"). InnerVolumeSpecName "kube-api-access-8skqx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.961175 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ba340e8e-9795-40f8-9690-c224b952b5e6" (UID: "ba340e8e-9795-40f8-9690-c224b952b5e6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.989738 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.989802 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.989851 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.990443 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.990550 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26" gracePeriod=600 Dec 03 17:16:48 crc kubenswrapper[4787]: I1203 17:16:48.991601 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8c938fa3-8ab3-4d20-b9ef-2fa192313e98" (UID: "8c938fa3-8ab3-4d20-b9ef-2fa192313e98"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.041456 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-72cww\" (UniqueName: \"kubernetes.io/projected/ba340e8e-9795-40f8-9690-c224b952b5e6-kube-api-access-72cww\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.041501 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.041514 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.041526 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ba340e8e-9795-40f8-9690-c224b952b5e6-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.041539 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8skqx\" (UniqueName: \"kubernetes.io/projected/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-kube-api-access-8skqx\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.041551 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8c938fa3-8ab3-4d20-b9ef-2fa192313e98-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.381166 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hc88l" event={"ID":"8c938fa3-8ab3-4d20-b9ef-2fa192313e98","Type":"ContainerDied","Data":"12a5ba350c2090eb34ec9b60f0ab5e91be18f98fc700dbe2ac37ec65b1e9fcaf"} Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.381232 4787 scope.go:117] "RemoveContainer" containerID="16f8f7b3c41a08f81c78dbc873f9713ee9864e61abff7bf7e2650c3b45442477" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.381351 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hc88l" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.385914 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-llckd" event={"ID":"ba340e8e-9795-40f8-9690-c224b952b5e6","Type":"ContainerDied","Data":"19e4e6de7a632500619945ea24b0c30b570f325999e629245f968c5e1a7b9e36"} Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.386049 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-llckd" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.401984 4787 scope.go:117] "RemoveContainer" containerID="193dbbb09ecba3dd05c9a03a2a01e7a3499d3b86533c1c4e8367caf4fa8e6e56" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.420337 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hc88l"] Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.423068 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hc88l"] Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.430615 4787 scope.go:117] "RemoveContainer" containerID="57481fa2e2f33f5fa59aeb643105d77d00bb11aaa692f12b1ccd483e8e4e0334" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.449468 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-llckd"] Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.450000 4787 scope.go:117] "RemoveContainer" containerID="066f0274138220a0a7fae618b82a115d07c86427f738ed458e4506000995a7b1" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.452565 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-llckd"] Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.460401 4787 scope.go:117] "RemoveContainer" containerID="513d315f6926494d40aad1a45794fe2296ba33dd3cce90660437f93fa754f75a" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.473706 4787 scope.go:117] "RemoveContainer" containerID="d004064d11226c05b0c220865c217bd0c2ac9f1993ed5036f62fb9f01a8d9f30" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.772332 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" path="/var/lib/kubelet/pods/4435fc6e-cc89-4399-89dc-c2a87c9b7153/volumes" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.773215 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" path="/var/lib/kubelet/pods/8c938fa3-8ab3-4d20-b9ef-2fa192313e98/volumes" Dec 03 17:16:49 crc kubenswrapper[4787]: I1203 17:16:49.773757 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba340e8e-9795-40f8-9690-c224b952b5e6" path="/var/lib/kubelet/pods/ba340e8e-9795-40f8-9690-c224b952b5e6/volumes" Dec 03 17:16:50 crc kubenswrapper[4787]: I1203 17:16:50.395049 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54jtg" event={"ID":"2b45ac06-f7eb-47b8-b1c8-af110a7da844","Type":"ContainerStarted","Data":"d49d37b6a018ec45695946ec36a92ba03a7ff7a992081fa39a04352a2f3ea43b"} Dec 03 17:16:50 crc kubenswrapper[4787]: I1203 17:16:50.398526 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26" exitCode=0 Dec 03 17:16:50 crc kubenswrapper[4787]: I1203 17:16:50.398610 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26"} Dec 03 17:16:50 crc kubenswrapper[4787]: I1203 17:16:50.416321 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-54jtg" podStartSLOduration=4.558913781 podStartE2EDuration="1m9.416303034s" podCreationTimestamp="2025-12-03 17:15:41 +0000 UTC" firstStartedPulling="2025-12-03 17:15:43.63895124 +0000 UTC m=+160.456422189" lastFinishedPulling="2025-12-03 17:16:48.496340483 +0000 UTC m=+225.313811442" observedRunningTime="2025-12-03 17:16:50.412898697 +0000 UTC m=+227.230369666" watchObservedRunningTime="2025-12-03 17:16:50.416303034 +0000 UTC m=+227.233774013" Dec 03 17:16:50 crc kubenswrapper[4787]: I1203 17:16:50.441990 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-bhg5b"] Dec 03 17:16:51 crc kubenswrapper[4787]: I1203 17:16:51.407609 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"d126a06f7bcbd1acb6bb4b36067339af8c76b0636d70dda2e29f3ff99ca18e1c"} Dec 03 17:16:51 crc kubenswrapper[4787]: I1203 17:16:51.451976 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:16:51 crc kubenswrapper[4787]: I1203 17:16:51.452097 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:16:51 crc kubenswrapper[4787]: I1203 17:16:51.496577 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:16:53 crc kubenswrapper[4787]: I1203 17:16:53.107064 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:16:53 crc kubenswrapper[4787]: I1203 17:16:53.107110 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:16:53 crc kubenswrapper[4787]: I1203 17:16:53.140509 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:16:53 crc kubenswrapper[4787]: I1203 17:16:53.473828 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:16:55 crc kubenswrapper[4787]: I1203 17:16:55.984195 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lg989"] Dec 03 17:16:55 crc kubenswrapper[4787]: I1203 17:16:55.984950 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lg989" podUID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" containerName="registry-server" containerID="cri-o://58eef3295e8300e4ba409fd5dab42843184602c96e3536f3042d95e8dc9295e6" gracePeriod=2 Dec 03 17:16:56 crc kubenswrapper[4787]: I1203 17:16:56.438293 4787 generic.go:334] "Generic (PLEG): container finished" podID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" containerID="58eef3295e8300e4ba409fd5dab42843184602c96e3536f3042d95e8dc9295e6" exitCode=0 Dec 03 17:16:56 crc kubenswrapper[4787]: I1203 17:16:56.438364 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lg989" event={"ID":"a7b5ebba-5b07-430f-90ea-282fe60f6e5a","Type":"ContainerDied","Data":"58eef3295e8300e4ba409fd5dab42843184602c96e3536f3042d95e8dc9295e6"} Dec 03 17:16:56 crc kubenswrapper[4787]: I1203 17:16:56.845808 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:16:56 crc kubenswrapper[4787]: I1203 17:16:56.940806 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-catalog-content\") pod \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " Dec 03 17:16:56 crc kubenswrapper[4787]: I1203 17:16:56.940918 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fnt5f\" (UniqueName: \"kubernetes.io/projected/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-kube-api-access-fnt5f\") pod \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " Dec 03 17:16:56 crc kubenswrapper[4787]: I1203 17:16:56.940950 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-utilities\") pod \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\" (UID: \"a7b5ebba-5b07-430f-90ea-282fe60f6e5a\") " Dec 03 17:16:56 crc kubenswrapper[4787]: I1203 17:16:56.942107 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-utilities" (OuterVolumeSpecName: "utilities") pod "a7b5ebba-5b07-430f-90ea-282fe60f6e5a" (UID: "a7b5ebba-5b07-430f-90ea-282fe60f6e5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:16:56 crc kubenswrapper[4787]: I1203 17:16:56.946455 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-kube-api-access-fnt5f" (OuterVolumeSpecName: "kube-api-access-fnt5f") pod "a7b5ebba-5b07-430f-90ea-282fe60f6e5a" (UID: "a7b5ebba-5b07-430f-90ea-282fe60f6e5a"). InnerVolumeSpecName "kube-api-access-fnt5f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.042970 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fnt5f\" (UniqueName: \"kubernetes.io/projected/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-kube-api-access-fnt5f\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.043054 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.057417 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a7b5ebba-5b07-430f-90ea-282fe60f6e5a" (UID: "a7b5ebba-5b07-430f-90ea-282fe60f6e5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.144233 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a7b5ebba-5b07-430f-90ea-282fe60f6e5a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.447634 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lg989" event={"ID":"a7b5ebba-5b07-430f-90ea-282fe60f6e5a","Type":"ContainerDied","Data":"fe65942452f1817fcd636f9133c91a1f2a1b9523a54364c8ba2f33c22ca728c1"} Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.447696 4787 scope.go:117] "RemoveContainer" containerID="58eef3295e8300e4ba409fd5dab42843184602c96e3536f3042d95e8dc9295e6" Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.447824 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lg989" Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.465401 4787 scope.go:117] "RemoveContainer" containerID="5c29234e8ce390ad2ae3dfc60ffb3110c233d0c8228e8d07875f273b8927d6c7" Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.474960 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lg989"] Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.485448 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lg989"] Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.485467 4787 scope.go:117] "RemoveContainer" containerID="6d62ed84a9e0f2311f39cba1fe0dc3f060487b2c49f1127b33c4ec49310ce6ec" Dec 03 17:16:57 crc kubenswrapper[4787]: I1203 17:16:57.772879 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" path="/var/lib/kubelet/pods/a7b5ebba-5b07-430f-90ea-282fe60f6e5a/volumes" Dec 03 17:17:01 crc kubenswrapper[4787]: I1203 17:17:01.498119 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.214822 4787 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.217849 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" containerName="extract-content" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.217885 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" containerName="extract-content" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.217903 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.217916 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.217931 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba340e8e-9795-40f8-9690-c224b952b5e6" containerName="extract-content" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.217943 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba340e8e-9795-40f8-9690-c224b952b5e6" containerName="extract-content" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.217962 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" containerName="extract-content" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.217973 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" containerName="extract-content" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.217989 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" containerName="extract-utilities" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218000 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" containerName="extract-utilities" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.218037 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218049 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.218064 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218075 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.218091 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba340e8e-9795-40f8-9690-c224b952b5e6" containerName="extract-utilities" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218102 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba340e8e-9795-40f8-9690-c224b952b5e6" containerName="extract-utilities" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.218123 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" containerName="extract-utilities" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218134 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" containerName="extract-utilities" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.218151 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" containerName="extract-content" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218163 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" containerName="extract-content" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.218179 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba340e8e-9795-40f8-9690-c224b952b5e6" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218191 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba340e8e-9795-40f8-9690-c224b952b5e6" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.218205 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" containerName="extract-utilities" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218217 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" containerName="extract-utilities" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218393 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8c938fa3-8ab3-4d20-b9ef-2fa192313e98" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218417 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4435fc6e-cc89-4399-89dc-c2a87c9b7153" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218432 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a7b5ebba-5b07-430f-90ea-282fe60f6e5a" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218449 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba340e8e-9795-40f8-9690-c224b952b5e6" containerName="registry-server" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218909 4787 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.218945 4787 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219098 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.219152 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219172 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.219192 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219205 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.219242 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219253 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.219276 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219287 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.219301 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219312 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.219330 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219341 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 17:17:02 crc kubenswrapper[4787]: E1203 17:17:02.219356 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219366 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219509 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219525 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219544 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219561 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219573 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.219588 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.220588 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49" gracePeriod=15 Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.220611 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927" gracePeriod=15 Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.220682 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9" gracePeriod=15 Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.220732 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a" gracePeriod=15 Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.220862 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d" gracePeriod=15 Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.223373 4787 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="f4b27818a5e8e43d0dc095d08835c792" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.307761 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.307815 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.307839 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.307858 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.307875 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.308092 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.308185 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.308254 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.409908 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.410162 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.410352 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.410585 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.410685 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.410099 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.410295 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.410641 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.410489 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.410866 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.410908 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.411044 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.411197 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.411366 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.411329 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.411478 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.495977 4787 generic.go:334] "Generic (PLEG): container finished" podID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" containerID="615ce9f7de0380159065e5ecc8d1f573f0d7a38d58733fc2e437263d87a91ae2" exitCode=0 Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.496320 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2cfbeef7-cc82-419e-ae28-c6eef3202ebc","Type":"ContainerDied","Data":"615ce9f7de0380159065e5ecc8d1f573f0d7a38d58733fc2e437263d87a91ae2"} Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.497112 4787 status_manager.go:851] "Failed to get status for pod" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.499063 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.501151 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.501808 4787 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927" exitCode=0 Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.501888 4787 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a" exitCode=0 Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.501965 4787 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9" exitCode=0 Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.502065 4787 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d" exitCode=2 Dec 03 17:17:02 crc kubenswrapper[4787]: I1203 17:17:02.502145 4787 scope.go:117] "RemoveContainer" containerID="13b43fef724a4649c874daec8ab51a14d344550ce45c752d5643ffdb0fef98fb" Dec 03 17:17:03 crc kubenswrapper[4787]: I1203 17:17:03.513915 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 17:17:03 crc kubenswrapper[4787]: I1203 17:17:03.768572 4787 status_manager.go:851] "Failed to get status for pod" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:03 crc kubenswrapper[4787]: I1203 17:17:03.776399 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:17:03 crc kubenswrapper[4787]: I1203 17:17:03.776887 4787 status_manager.go:851] "Failed to get status for pod" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:03 crc kubenswrapper[4787]: I1203 17:17:03.932768 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kube-api-access\") pod \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " Dec 03 17:17:03 crc kubenswrapper[4787]: I1203 17:17:03.933038 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-var-lock\") pod \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " Dec 03 17:17:03 crc kubenswrapper[4787]: I1203 17:17:03.933075 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kubelet-dir\") pod \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\" (UID: \"2cfbeef7-cc82-419e-ae28-c6eef3202ebc\") " Dec 03 17:17:03 crc kubenswrapper[4787]: I1203 17:17:03.933320 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-var-lock" (OuterVolumeSpecName: "var-lock") pod "2cfbeef7-cc82-419e-ae28-c6eef3202ebc" (UID: "2cfbeef7-cc82-419e-ae28-c6eef3202ebc"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:03 crc kubenswrapper[4787]: I1203 17:17:03.933439 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "2cfbeef7-cc82-419e-ae28-c6eef3202ebc" (UID: "2cfbeef7-cc82-419e-ae28-c6eef3202ebc"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:03 crc kubenswrapper[4787]: I1203 17:17:03.939463 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "2cfbeef7-cc82-419e-ae28-c6eef3202ebc" (UID: "2cfbeef7-cc82-419e-ae28-c6eef3202ebc"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.034387 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.034429 4787 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.034441 4787 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/2cfbeef7-cc82-419e-ae28-c6eef3202ebc-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.522407 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"2cfbeef7-cc82-419e-ae28-c6eef3202ebc","Type":"ContainerDied","Data":"262421ddf72cca4d12499646acb1eca141f3117d75bef12aa683ff1df2251f00"} Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.522662 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="262421ddf72cca4d12499646acb1eca141f3117d75bef12aa683ff1df2251f00" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.522491 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.588581 4787 status_manager.go:851] "Failed to get status for pod" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.592702 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.593404 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.593856 4787 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.594173 4787 status_manager.go:851] "Failed to get status for pod" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.747689 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.747854 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.747880 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.747911 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.747980 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.748070 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.748406 4787 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.748446 4787 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:04 crc kubenswrapper[4787]: I1203 17:17:04.748465 4787 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.528994 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.529439 4787 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49" exitCode=0 Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.529478 4787 scope.go:117] "RemoveContainer" containerID="c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.529582 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.543177 4787 status_manager.go:851] "Failed to get status for pod" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.543740 4787 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.549090 4787 scope.go:117] "RemoveContainer" containerID="7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.562528 4787 scope.go:117] "RemoveContainer" containerID="33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.580793 4787 scope.go:117] "RemoveContainer" containerID="8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.593880 4787 scope.go:117] "RemoveContainer" containerID="08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.607961 4787 scope.go:117] "RemoveContainer" containerID="1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.628656 4787 scope.go:117] "RemoveContainer" containerID="c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.630989 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\": container with ID starting with c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927 not found: ID does not exist" containerID="c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.631047 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927"} err="failed to get container status \"c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\": rpc error: code = NotFound desc = could not find container \"c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927\": container with ID starting with c221f59bb92e34f57db608f823e2e17ef4b20821ea60c285f6af9347b4645927 not found: ID does not exist" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.631075 4787 scope.go:117] "RemoveContainer" containerID="7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.631389 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\": container with ID starting with 7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a not found: ID does not exist" containerID="7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.631424 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a"} err="failed to get container status \"7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\": rpc error: code = NotFound desc = could not find container \"7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a\": container with ID starting with 7337c76255073dfe9fde1e828b46cb827b985c17333b941576bd988cbc01d82a not found: ID does not exist" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.631445 4787 scope.go:117] "RemoveContainer" containerID="33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.631799 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\": container with ID starting with 33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9 not found: ID does not exist" containerID="33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.631827 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9"} err="failed to get container status \"33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\": rpc error: code = NotFound desc = could not find container \"33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9\": container with ID starting with 33910da057ea932a8374d632d88708c95774e4308839c555ccbd1b38ff8573e9 not found: ID does not exist" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.631844 4787 scope.go:117] "RemoveContainer" containerID="8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.632162 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\": container with ID starting with 8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d not found: ID does not exist" containerID="8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.632830 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d"} err="failed to get container status \"8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\": rpc error: code = NotFound desc = could not find container \"8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d\": container with ID starting with 8baf317f045b5386bcdb77f48c611056d6c5388d91f87197c741d5b70e757f9d not found: ID does not exist" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.632892 4787 scope.go:117] "RemoveContainer" containerID="08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.633282 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\": container with ID starting with 08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49 not found: ID does not exist" containerID="08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.633306 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49"} err="failed to get container status \"08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\": rpc error: code = NotFound desc = could not find container \"08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49\": container with ID starting with 08ee0c1980e8f0fcc984bf67e8b7a4979731b1a397bca7be6aa3e725cd725d49 not found: ID does not exist" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.633322 4787 scope.go:117] "RemoveContainer" containerID="1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.633708 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\": container with ID starting with 1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375 not found: ID does not exist" containerID="1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.633731 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375"} err="failed to get container status \"1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\": rpc error: code = NotFound desc = could not find container \"1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375\": container with ID starting with 1fa2f6b73a4bf32e7ad0698c17ec47d41e0b41ef300f4f17957f10dcc0f2b375 not found: ID does not exist" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.772229 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.933678 4787 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.933924 4787 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.934168 4787 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.934354 4787 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.934590 4787 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:05 crc kubenswrapper[4787]: I1203 17:17:05.934614 4787 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 03 17:17:05 crc kubenswrapper[4787]: E1203 17:17:05.934794 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="200ms" Dec 03 17:17:06 crc kubenswrapper[4787]: E1203 17:17:06.135999 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="400ms" Dec 03 17:17:06 crc kubenswrapper[4787]: E1203 17:17:06.537642 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="800ms" Dec 03 17:17:07 crc kubenswrapper[4787]: E1203 17:17:07.269233 4787 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.65:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:07 crc kubenswrapper[4787]: I1203 17:17:07.269648 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:07 crc kubenswrapper[4787]: W1203 17:17:07.286878 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-3ee210c1dbb775f8ff67f558691220b3214fe429517bc949723f057ca0b2e282 WatchSource:0}: Error finding container 3ee210c1dbb775f8ff67f558691220b3214fe429517bc949723f057ca0b2e282: Status 404 returned error can't find the container with id 3ee210c1dbb775f8ff67f558691220b3214fe429517bc949723f057ca0b2e282 Dec 03 17:17:07 crc kubenswrapper[4787]: E1203 17:17:07.290476 4787 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.65:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dc415e38b2f47 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 17:17:07.290009415 +0000 UTC m=+244.107480364,LastTimestamp:2025-12-03 17:17:07.290009415 +0000 UTC m=+244.107480364,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 17:17:07 crc kubenswrapper[4787]: E1203 17:17:07.338569 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="1.6s" Dec 03 17:17:07 crc kubenswrapper[4787]: I1203 17:17:07.540853 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5"} Dec 03 17:17:07 crc kubenswrapper[4787]: I1203 17:17:07.540907 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"3ee210c1dbb775f8ff67f558691220b3214fe429517bc949723f057ca0b2e282"} Dec 03 17:17:07 crc kubenswrapper[4787]: E1203 17:17:07.541564 4787 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.65:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:07 crc kubenswrapper[4787]: I1203 17:17:07.541704 4787 status_manager.go:851] "Failed to get status for pod" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:08 crc kubenswrapper[4787]: E1203 17:17:08.940396 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="3.2s" Dec 03 17:17:10 crc kubenswrapper[4787]: E1203 17:17:10.575308 4787 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.65:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187dc415e38b2f47 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-03 17:17:07.290009415 +0000 UTC m=+244.107480364,LastTimestamp:2025-12-03 17:17:07.290009415 +0000 UTC m=+244.107480364,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 03 17:17:12 crc kubenswrapper[4787]: E1203 17:17:12.141870 4787 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.65:6443: connect: connection refused" interval="6.4s" Dec 03 17:17:13 crc kubenswrapper[4787]: I1203 17:17:13.767813 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:13 crc kubenswrapper[4787]: I1203 17:17:13.767918 4787 status_manager.go:851] "Failed to get status for pod" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:13 crc kubenswrapper[4787]: I1203 17:17:13.768575 4787 status_manager.go:851] "Failed to get status for pod" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:13 crc kubenswrapper[4787]: I1203 17:17:13.783227 4787 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:13 crc kubenswrapper[4787]: I1203 17:17:13.783496 4787 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:13 crc kubenswrapper[4787]: E1203 17:17:13.783995 4787 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:13 crc kubenswrapper[4787]: I1203 17:17:13.784448 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:14 crc kubenswrapper[4787]: I1203 17:17:14.586510 4787 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="0c3f079ca5ad7fcbe82b0ff35c9130d98689e687d7bfa073e5f6f3a815a5e958" exitCode=0 Dec 03 17:17:14 crc kubenswrapper[4787]: I1203 17:17:14.586691 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"0c3f079ca5ad7fcbe82b0ff35c9130d98689e687d7bfa073e5f6f3a815a5e958"} Dec 03 17:17:14 crc kubenswrapper[4787]: I1203 17:17:14.587084 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"0d4643b2373758b3273c30714b7978268bafc4038088b38aecbb07e9ec598248"} Dec 03 17:17:14 crc kubenswrapper[4787]: I1203 17:17:14.587539 4787 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:14 crc kubenswrapper[4787]: I1203 17:17:14.587578 4787 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:14 crc kubenswrapper[4787]: I1203 17:17:14.588233 4787 status_manager.go:851] "Failed to get status for pod" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" Dec 03 17:17:14 crc kubenswrapper[4787]: E1203 17:17:14.588269 4787 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.65:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.468558 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" podUID="e19b19ce-efa6-4d43-af6c-bb6aed7c232d" containerName="oauth-openshift" containerID="cri-o://24cf7301af91405740dd42a3d2da7c37a452748aed65f4f10b6c03138839da9b" gracePeriod=15 Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.594550 4787 generic.go:334] "Generic (PLEG): container finished" podID="e19b19ce-efa6-4d43-af6c-bb6aed7c232d" containerID="24cf7301af91405740dd42a3d2da7c37a452748aed65f4f10b6c03138839da9b" exitCode=0 Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.594640 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" event={"ID":"e19b19ce-efa6-4d43-af6c-bb6aed7c232d","Type":"ContainerDied","Data":"24cf7301af91405740dd42a3d2da7c37a452748aed65f4f10b6c03138839da9b"} Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.597167 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"322bae9e60090beeb46ba65d6e3e631848033f6cfa796062f66d6f448df4fd31"} Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.597223 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"150bfe4b8029b0246e233afddeb45cf8b8f7904344162cba44462017141f3ae7"} Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.597235 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3090b4dd2da5544a01e75cf0963bbe3becdef9ee97328448d4770a542805f014"} Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.597245 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"e17b78d8440c02c85adf451ccb1601e66353fd8f3481221c3c561dbb7fa12d03"} Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.836048 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986294 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-ocp-branding-template\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986343 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-dir\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986373 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-cliconfig\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986395 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-provider-selection\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986426 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-idp-0-file-data\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986450 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-service-ca\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986473 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986491 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-router-certs\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986589 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-error\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986632 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-policies\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986701 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-serving-cert\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986730 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-trusted-ca-bundle\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986814 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xrzg\" (UniqueName: \"kubernetes.io/projected/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-kube-api-access-8xrzg\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986867 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-session\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.986891 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-login\") pod \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\" (UID: \"e19b19ce-efa6-4d43-af6c-bb6aed7c232d\") " Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.987618 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.987643 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.987675 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.988009 4787 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.988056 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.992739 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.992817 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-kube-api-access-8xrzg" (OuterVolumeSpecName: "kube-api-access-8xrzg") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "kube-api-access-8xrzg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.993534 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.993871 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:17:15 crc kubenswrapper[4787]: I1203 17:17:15.995345 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.002314 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.002686 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.007259 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.012484 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "e19b19ce-efa6-4d43-af6c-bb6aed7c232d" (UID: "e19b19ce-efa6-4d43-af6c-bb6aed7c232d"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089531 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089575 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089588 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089603 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089617 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089628 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089642 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089657 4787 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089670 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089683 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089694 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xrzg\" (UniqueName: \"kubernetes.io/projected/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-kube-api-access-8xrzg\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089705 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.089715 4787 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/e19b19ce-efa6-4d43-af6c-bb6aed7c232d-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.609490 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"08e8084a4d29c6f5878bf7c5412dfc8c340820535d9b68defc6a46459002bb05"} Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.610554 4787 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.610601 4787 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.612245 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" event={"ID":"e19b19ce-efa6-4d43-af6c-bb6aed7c232d","Type":"ContainerDied","Data":"f778721fe4300af26b049636592d0f61b4792714f47688c6e65026e89ce6a9d9"} Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.612284 4787 scope.go:117] "RemoveContainer" containerID="24cf7301af91405740dd42a3d2da7c37a452748aed65f4f10b6c03138839da9b" Dec 03 17:17:16 crc kubenswrapper[4787]: I1203 17:17:16.612336 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-bhg5b" Dec 03 17:17:16 crc kubenswrapper[4787]: E1203 17:17:16.857352 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-conmon-79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:17:17 crc kubenswrapper[4787]: I1203 17:17:17.625685 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 17:17:17 crc kubenswrapper[4787]: I1203 17:17:17.626963 4787 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb" exitCode=1 Dec 03 17:17:17 crc kubenswrapper[4787]: I1203 17:17:17.627106 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb"} Dec 03 17:17:17 crc kubenswrapper[4787]: I1203 17:17:17.628358 4787 scope.go:117] "RemoveContainer" containerID="79d38ff37a25598f0e1224bb306c86e0563f8e0252f209578d293622f791dadb" Dec 03 17:17:18 crc kubenswrapper[4787]: I1203 17:17:18.432727 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:17:18 crc kubenswrapper[4787]: I1203 17:17:18.644003 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 03 17:17:18 crc kubenswrapper[4787]: I1203 17:17:18.644160 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"91b3522d6ce7ce4175fccca178536496ba2ec363a98ff4c22de5550a56bb22f7"} Dec 03 17:17:18 crc kubenswrapper[4787]: I1203 17:17:18.785325 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:18 crc kubenswrapper[4787]: I1203 17:17:18.785630 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:18 crc kubenswrapper[4787]: I1203 17:17:18.791128 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:19 crc kubenswrapper[4787]: I1203 17:17:19.493486 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:17:19 crc kubenswrapper[4787]: I1203 17:17:19.500351 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:17:19 crc kubenswrapper[4787]: I1203 17:17:19.650292 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:17:21 crc kubenswrapper[4787]: I1203 17:17:21.620487 4787 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:21 crc kubenswrapper[4787]: I1203 17:17:21.659742 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:21 crc kubenswrapper[4787]: I1203 17:17:21.659780 4787 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:21 crc kubenswrapper[4787]: I1203 17:17:21.659807 4787 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:21 crc kubenswrapper[4787]: I1203 17:17:21.664122 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:22 crc kubenswrapper[4787]: I1203 17:17:22.664507 4787 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:22 crc kubenswrapper[4787]: I1203 17:17:22.665242 4787 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:23 crc kubenswrapper[4787]: I1203 17:17:23.669892 4787 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:23 crc kubenswrapper[4787]: I1203 17:17:23.670244 4787 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="7a92940f-2d9e-466f-8cfd-5a3cf91740f6" Dec 03 17:17:23 crc kubenswrapper[4787]: I1203 17:17:23.784932 4787 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="c872c46c-084d-4ff0-97ac-b94cf24ff002" Dec 03 17:17:27 crc kubenswrapper[4787]: I1203 17:17:27.754686 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 03 17:17:32 crc kubenswrapper[4787]: I1203 17:17:32.735909 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 03 17:17:33 crc kubenswrapper[4787]: I1203 17:17:33.152154 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 03 17:17:33 crc kubenswrapper[4787]: I1203 17:17:33.421386 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 03 17:17:33 crc kubenswrapper[4787]: I1203 17:17:33.708778 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 03 17:17:33 crc kubenswrapper[4787]: I1203 17:17:33.762757 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 03 17:17:33 crc kubenswrapper[4787]: I1203 17:17:33.791474 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 17:17:33 crc kubenswrapper[4787]: I1203 17:17:33.919787 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 03 17:17:33 crc kubenswrapper[4787]: I1203 17:17:33.951561 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.096892 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.123797 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.265340 4787 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.368506 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.371177 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.383045 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.422863 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.449160 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.506805 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.525903 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.550420 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 03 17:17:34 crc kubenswrapper[4787]: I1203 17:17:34.950852 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 17:17:35 crc kubenswrapper[4787]: I1203 17:17:35.031603 4787 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 03 17:17:35 crc kubenswrapper[4787]: I1203 17:17:35.057684 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 03 17:17:35 crc kubenswrapper[4787]: I1203 17:17:35.253451 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 03 17:17:35 crc kubenswrapper[4787]: I1203 17:17:35.385986 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 03 17:17:35 crc kubenswrapper[4787]: I1203 17:17:35.416958 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 03 17:17:35 crc kubenswrapper[4787]: I1203 17:17:35.630149 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 03 17:17:35 crc kubenswrapper[4787]: I1203 17:17:35.667804 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 03 17:17:35 crc kubenswrapper[4787]: I1203 17:17:35.803606 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 03 17:17:35 crc kubenswrapper[4787]: I1203 17:17:35.827886 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 03 17:17:35 crc kubenswrapper[4787]: I1203 17:17:35.979384 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.023438 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.054845 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.099901 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.148158 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.241835 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.396744 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.412883 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.515111 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.552073 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.652526 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.849440 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.871349 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.889674 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.915465 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 03 17:17:36 crc kubenswrapper[4787]: I1203 17:17:36.917606 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 03 17:17:37 crc kubenswrapper[4787]: I1203 17:17:37.469467 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 03 17:17:37 crc kubenswrapper[4787]: I1203 17:17:37.473820 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 03 17:17:37 crc kubenswrapper[4787]: I1203 17:17:37.501481 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 03 17:17:37 crc kubenswrapper[4787]: I1203 17:17:37.560908 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 03 17:17:37 crc kubenswrapper[4787]: I1203 17:17:37.678837 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 03 17:17:37 crc kubenswrapper[4787]: I1203 17:17:37.701778 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 17:17:37 crc kubenswrapper[4787]: I1203 17:17:37.718588 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 03 17:17:37 crc kubenswrapper[4787]: I1203 17:17:37.726199 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 03 17:17:37 crc kubenswrapper[4787]: I1203 17:17:37.730153 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 03 17:17:37 crc kubenswrapper[4787]: I1203 17:17:37.811326 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 03 17:17:37 crc kubenswrapper[4787]: I1203 17:17:37.831715 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.016211 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.058076 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.088860 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.116970 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.274421 4787 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.327804 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.358507 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.393567 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.582010 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.584910 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.815381 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.886289 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.889277 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.942503 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 03 17:17:38 crc kubenswrapper[4787]: I1203 17:17:38.990957 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.064534 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.082150 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.162345 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.195539 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.265430 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.297087 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.311008 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.311766 4787 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.337717 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.379452 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.392804 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.413971 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.499851 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.527689 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.546959 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.587538 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.634220 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.643313 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.670998 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.726090 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.763424 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.866247 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.911380 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 03 17:17:39 crc kubenswrapper[4787]: I1203 17:17:39.938728 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 03 17:17:40 crc kubenswrapper[4787]: I1203 17:17:40.197656 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 03 17:17:40 crc kubenswrapper[4787]: I1203 17:17:40.202548 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 03 17:17:40 crc kubenswrapper[4787]: I1203 17:17:40.578241 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 03 17:17:40 crc kubenswrapper[4787]: I1203 17:17:40.608183 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 03 17:17:40 crc kubenswrapper[4787]: I1203 17:17:40.633162 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 03 17:17:40 crc kubenswrapper[4787]: I1203 17:17:40.735936 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 03 17:17:40 crc kubenswrapper[4787]: I1203 17:17:40.801387 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 03 17:17:40 crc kubenswrapper[4787]: I1203 17:17:40.896667 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 03 17:17:40 crc kubenswrapper[4787]: I1203 17:17:40.911915 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.002468 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.026867 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.084091 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.105909 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.210984 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.212413 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.248672 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.337131 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.349719 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.360292 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.362053 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.407628 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.537656 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.556913 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.582618 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.615597 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.644279 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.678072 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.678758 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.746096 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.790037 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.813276 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.885974 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 03 17:17:41 crc kubenswrapper[4787]: I1203 17:17:41.917092 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.009549 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.060990 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.061136 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.086223 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.126133 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.160227 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.224633 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.326225 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.353664 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.373886 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.511709 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.577824 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.588828 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.652442 4787 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.654208 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.686636 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.696380 4787 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.701693 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-bhg5b"] Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.701759 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.707566 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.721817 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=21.721802489 podStartE2EDuration="21.721802489s" podCreationTimestamp="2025-12-03 17:17:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:17:42.72150522 +0000 UTC m=+279.538976209" watchObservedRunningTime="2025-12-03 17:17:42.721802489 +0000 UTC m=+279.539273448" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.744881 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.788361 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.862365 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.893244 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.961758 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.968096 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.968995 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.969047 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 03 17:17:42 crc kubenswrapper[4787]: I1203 17:17:42.995600 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.071676 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.126522 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.126536 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.153348 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6cb668d466-h8rf9"] Dec 03 17:17:43 crc kubenswrapper[4787]: E1203 17:17:43.153537 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e19b19ce-efa6-4d43-af6c-bb6aed7c232d" containerName="oauth-openshift" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.153550 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e19b19ce-efa6-4d43-af6c-bb6aed7c232d" containerName="oauth-openshift" Dec 03 17:17:43 crc kubenswrapper[4787]: E1203 17:17:43.153571 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" containerName="installer" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.153577 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" containerName="installer" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.153658 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2cfbeef7-cc82-419e-ae28-c6eef3202ebc" containerName="installer" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.153671 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e19b19ce-efa6-4d43-af6c-bb6aed7c232d" containerName="oauth-openshift" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.154002 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.156243 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.156385 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.156395 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.157088 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.157175 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.157407 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.157488 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.157413 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.157998 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.158099 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.158653 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.158691 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.166335 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.167502 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.174397 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.175579 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.207972 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.222596 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.222650 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.222686 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-audit-policies\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.222718 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-session\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.222807 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.222842 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-template-error\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.222873 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.222900 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-template-login\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.222943 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.222997 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.223066 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtk8p\" (UniqueName: \"kubernetes.io/projected/852e8cbf-96dc-4a69-8b54-aed15de1396f-kube-api-access-vtk8p\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.223094 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.223145 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/852e8cbf-96dc-4a69-8b54-aed15de1396f-audit-dir\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.223195 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.283451 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.310279 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.319117 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.324953 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtk8p\" (UniqueName: \"kubernetes.io/projected/852e8cbf-96dc-4a69-8b54-aed15de1396f-kube-api-access-vtk8p\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325067 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325127 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/852e8cbf-96dc-4a69-8b54-aed15de1396f-audit-dir\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325169 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325224 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325252 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325299 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-audit-policies\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325322 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-session\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325386 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325421 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-template-error\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325478 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325504 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-template-login\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325558 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.325588 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.327482 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.327501 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-audit-policies\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.329645 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.331323 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.332855 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/852e8cbf-96dc-4a69-8b54-aed15de1396f-audit-dir\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.334717 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.338433 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-session\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.342498 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.343465 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.346635 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.347132 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-template-error\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.347862 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-template-login\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.355351 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/852e8cbf-96dc-4a69-8b54-aed15de1396f-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.367647 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtk8p\" (UniqueName: \"kubernetes.io/projected/852e8cbf-96dc-4a69-8b54-aed15de1396f-kube-api-access-vtk8p\") pod \"oauth-openshift-6cb668d466-h8rf9\" (UID: \"852e8cbf-96dc-4a69-8b54-aed15de1396f\") " pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.443641 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.475304 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.480154 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.480615 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.546316 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.547718 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.552327 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.621261 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.715649 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.731220 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.774964 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e19b19ce-efa6-4d43-af6c-bb6aed7c232d" path="/var/lib/kubelet/pods/e19b19ce-efa6-4d43-af6c-bb6aed7c232d/volumes" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.792979 4787 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.793550 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5" gracePeriod=5 Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.799847 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.838181 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.887441 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.916568 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 03 17:17:43 crc kubenswrapper[4787]: I1203 17:17:43.990879 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.017088 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.101144 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.135125 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.219887 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.245788 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.299008 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.326609 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.439266 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.593800 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.689802 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6cb668d466-h8rf9"] Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.715563 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.749898 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.797352 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.828625 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.847638 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.892433 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.910218 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6cb668d466-h8rf9"] Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.960227 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 03 17:17:44 crc kubenswrapper[4787]: I1203 17:17:44.961852 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.258063 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.316751 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.418243 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.440367 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.442270 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.497493 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.527804 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.813794 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" event={"ID":"852e8cbf-96dc-4a69-8b54-aed15de1396f","Type":"ContainerStarted","Data":"bdf5a9ecb5e4432b97b88699e29354f7602ac28f3d32ee26c2f4d9fe257fda49"} Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.813844 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" event={"ID":"852e8cbf-96dc-4a69-8b54-aed15de1396f","Type":"ContainerStarted","Data":"e905f8b3065d6c3325c916b17341d810de2f1237186fc760692f7a56283474a8"} Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.815102 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.823539 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.856908 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6cb668d466-h8rf9" podStartSLOduration=55.856839407 podStartE2EDuration="55.856839407s" podCreationTimestamp="2025-12-03 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:17:45.847249739 +0000 UTC m=+282.664720738" watchObservedRunningTime="2025-12-03 17:17:45.856839407 +0000 UTC m=+282.674310376" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.913286 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 03 17:17:45 crc kubenswrapper[4787]: I1203 17:17:45.972353 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 03 17:17:46 crc kubenswrapper[4787]: I1203 17:17:46.144119 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 03 17:17:46 crc kubenswrapper[4787]: I1203 17:17:46.179324 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 03 17:17:46 crc kubenswrapper[4787]: I1203 17:17:46.230475 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 03 17:17:46 crc kubenswrapper[4787]: I1203 17:17:46.269826 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 03 17:17:46 crc kubenswrapper[4787]: I1203 17:17:46.344094 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 03 17:17:46 crc kubenswrapper[4787]: I1203 17:17:46.534515 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 03 17:17:46 crc kubenswrapper[4787]: I1203 17:17:46.778835 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 03 17:17:46 crc kubenswrapper[4787]: I1203 17:17:46.948265 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 03 17:17:46 crc kubenswrapper[4787]: I1203 17:17:46.980480 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.230465 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.348954 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.383394 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.429367 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.452089 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.452174 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.456853 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.600729 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.757199 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.776592 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.803073 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.846457 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 03 17:17:47 crc kubenswrapper[4787]: I1203 17:17:47.916900 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 03 17:17:48 crc kubenswrapper[4787]: I1203 17:17:48.010724 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 03 17:17:48 crc kubenswrapper[4787]: I1203 17:17:48.047158 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 03 17:17:48 crc kubenswrapper[4787]: I1203 17:17:48.132890 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 03 17:17:48 crc kubenswrapper[4787]: I1203 17:17:48.434159 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 17:17:48 crc kubenswrapper[4787]: I1203 17:17:48.530003 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 03 17:17:48 crc kubenswrapper[4787]: I1203 17:17:48.785855 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.363459 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.363563 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.448175 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.496504 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.496551 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.496596 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.496627 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.496647 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.496816 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.497666 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.497710 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.497790 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.506884 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.597518 4787 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.597561 4787 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.597574 4787 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.597590 4787 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.597601 4787 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.645477 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.742373 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.774531 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.836218 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.836300 4787 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5" exitCode=137 Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.836376 4787 scope.go:117] "RemoveContainer" containerID="ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.836459 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.860057 4787 scope.go:117] "RemoveContainer" containerID="ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5" Dec 03 17:17:49 crc kubenswrapper[4787]: E1203 17:17:49.860752 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5\": container with ID starting with ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5 not found: ID does not exist" containerID="ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5" Dec 03 17:17:49 crc kubenswrapper[4787]: I1203 17:17:49.860859 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5"} err="failed to get container status \"ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5\": rpc error: code = NotFound desc = could not find container \"ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5\": container with ID starting with ef3cd6338c9f860ac3734b1a43f1ae13bbd03643f921e023f5f4f9bcf09cafc5 not found: ID does not exist" Dec 03 17:17:50 crc kubenswrapper[4787]: I1203 17:17:50.497446 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 03 17:17:50 crc kubenswrapper[4787]: I1203 17:17:50.807948 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 03 17:17:51 crc kubenswrapper[4787]: I1203 17:17:51.614796 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 03 17:17:51 crc kubenswrapper[4787]: I1203 17:17:51.688896 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 03 17:18:08 crc kubenswrapper[4787]: I1203 17:18:08.953304 4787 generic.go:334] "Generic (PLEG): container finished" podID="684133b3-732f-47c6-b15d-108c3e37397f" containerID="9cbe8ac34d758da238115f28c4c930ec9270479e51b277a9c2d76cba1e2ef25a" exitCode=0 Dec 03 17:18:08 crc kubenswrapper[4787]: I1203 17:18:08.953406 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" event={"ID":"684133b3-732f-47c6-b15d-108c3e37397f","Type":"ContainerDied","Data":"9cbe8ac34d758da238115f28c4c930ec9270479e51b277a9c2d76cba1e2ef25a"} Dec 03 17:18:08 crc kubenswrapper[4787]: I1203 17:18:08.955304 4787 scope.go:117] "RemoveContainer" containerID="9cbe8ac34d758da238115f28c4c930ec9270479e51b277a9c2d76cba1e2ef25a" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.025664 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-t88fj"] Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.026398 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" podUID="f41a2379-8c92-4d95-96cc-3d466bdcbc1d" containerName="controller-manager" containerID="cri-o://e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0" gracePeriod=30 Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.105135 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg"] Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.105585 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" podUID="670e0167-14c0-4547-9539-f303e18c33f3" containerName="route-controller-manager" containerID="cri-o://f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431" gracePeriod=30 Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.446716 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.490618 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.567669 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jlqrb\" (UniqueName: \"kubernetes.io/projected/670e0167-14c0-4547-9539-f303e18c33f3-kube-api-access-jlqrb\") pod \"670e0167-14c0-4547-9539-f303e18c33f3\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.567744 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-config\") pod \"670e0167-14c0-4547-9539-f303e18c33f3\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.567798 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-config\") pod \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.567824 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-client-ca\") pod \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.567851 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/670e0167-14c0-4547-9539-f303e18c33f3-serving-cert\") pod \"670e0167-14c0-4547-9539-f303e18c33f3\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.567880 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnp85\" (UniqueName: \"kubernetes.io/projected/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-kube-api-access-lnp85\") pod \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.567904 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-serving-cert\") pod \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.567983 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-client-ca\") pod \"670e0167-14c0-4547-9539-f303e18c33f3\" (UID: \"670e0167-14c0-4547-9539-f303e18c33f3\") " Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.568039 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-proxy-ca-bundles\") pod \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\" (UID: \"f41a2379-8c92-4d95-96cc-3d466bdcbc1d\") " Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.569312 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-client-ca" (OuterVolumeSpecName: "client-ca") pod "f41a2379-8c92-4d95-96cc-3d466bdcbc1d" (UID: "f41a2379-8c92-4d95-96cc-3d466bdcbc1d"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.569325 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "f41a2379-8c92-4d95-96cc-3d466bdcbc1d" (UID: "f41a2379-8c92-4d95-96cc-3d466bdcbc1d"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.569314 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-client-ca" (OuterVolumeSpecName: "client-ca") pod "670e0167-14c0-4547-9539-f303e18c33f3" (UID: "670e0167-14c0-4547-9539-f303e18c33f3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.569498 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-config" (OuterVolumeSpecName: "config") pod "670e0167-14c0-4547-9539-f303e18c33f3" (UID: "670e0167-14c0-4547-9539-f303e18c33f3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.569965 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-config" (OuterVolumeSpecName: "config") pod "f41a2379-8c92-4d95-96cc-3d466bdcbc1d" (UID: "f41a2379-8c92-4d95-96cc-3d466bdcbc1d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.574421 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/670e0167-14c0-4547-9539-f303e18c33f3-kube-api-access-jlqrb" (OuterVolumeSpecName: "kube-api-access-jlqrb") pod "670e0167-14c0-4547-9539-f303e18c33f3" (UID: "670e0167-14c0-4547-9539-f303e18c33f3"). InnerVolumeSpecName "kube-api-access-jlqrb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.574878 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/670e0167-14c0-4547-9539-f303e18c33f3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "670e0167-14c0-4547-9539-f303e18c33f3" (UID: "670e0167-14c0-4547-9539-f303e18c33f3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.574904 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "f41a2379-8c92-4d95-96cc-3d466bdcbc1d" (UID: "f41a2379-8c92-4d95-96cc-3d466bdcbc1d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.575962 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-kube-api-access-lnp85" (OuterVolumeSpecName: "kube-api-access-lnp85") pod "f41a2379-8c92-4d95-96cc-3d466bdcbc1d" (UID: "f41a2379-8c92-4d95-96cc-3d466bdcbc1d"). InnerVolumeSpecName "kube-api-access-lnp85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.670853 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.670909 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/670e0167-14c0-4547-9539-f303e18c33f3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.670932 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnp85\" (UniqueName: \"kubernetes.io/projected/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-kube-api-access-lnp85\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.670955 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.670979 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.670999 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.671046 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jlqrb\" (UniqueName: \"kubernetes.io/projected/670e0167-14c0-4547-9539-f303e18c33f3-kube-api-access-jlqrb\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.671068 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/670e0167-14c0-4547-9539-f303e18c33f3-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.671091 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f41a2379-8c92-4d95-96cc-3d466bdcbc1d-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.962897 4787 generic.go:334] "Generic (PLEG): container finished" podID="f41a2379-8c92-4d95-96cc-3d466bdcbc1d" containerID="e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0" exitCode=0 Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.963031 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" event={"ID":"f41a2379-8c92-4d95-96cc-3d466bdcbc1d","Type":"ContainerDied","Data":"e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0"} Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.963097 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" event={"ID":"f41a2379-8c92-4d95-96cc-3d466bdcbc1d","Type":"ContainerDied","Data":"1954648f6ee937697936af552aa349f74fa246b05c951b9479143bbdb161926b"} Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.963126 4787 scope.go:117] "RemoveContainer" containerID="e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.963291 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-t88fj" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.968070 4787 generic.go:334] "Generic (PLEG): container finished" podID="670e0167-14c0-4547-9539-f303e18c33f3" containerID="f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431" exitCode=0 Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.968170 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" event={"ID":"670e0167-14c0-4547-9539-f303e18c33f3","Type":"ContainerDied","Data":"f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431"} Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.968231 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" event={"ID":"670e0167-14c0-4547-9539-f303e18c33f3","Type":"ContainerDied","Data":"23ee204b10f5e8af6ca66522a09e820fb3039e17b6c10321e50a807949808c64"} Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.968261 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.972565 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" event={"ID":"684133b3-732f-47c6-b15d-108c3e37397f","Type":"ContainerStarted","Data":"83490f7b9a6f69c61f85d85df8bb788c0c19df6c6daf4cc3344c2204d85b4497"} Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.973064 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.976757 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.991293 4787 scope.go:117] "RemoveContainer" containerID="e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0" Dec 03 17:18:09 crc kubenswrapper[4787]: E1203 17:18:09.991893 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0\": container with ID starting with e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0 not found: ID does not exist" containerID="e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.991954 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0"} err="failed to get container status \"e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0\": rpc error: code = NotFound desc = could not find container \"e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0\": container with ID starting with e147a7c84f2d65470bcec6fd5464a8f3824583d94d7788342fe42cda3ad2ffe0 not found: ID does not exist" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.991993 4787 scope.go:117] "RemoveContainer" containerID="f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431" Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.992134 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-t88fj"] Dec 03 17:18:09 crc kubenswrapper[4787]: I1203 17:18:09.999292 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-t88fj"] Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.027445 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg"] Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.030995 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-m4lbg"] Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.039071 4787 scope.go:117] "RemoveContainer" containerID="f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431" Dec 03 17:18:10 crc kubenswrapper[4787]: E1203 17:18:10.039737 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431\": container with ID starting with f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431 not found: ID does not exist" containerID="f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.039767 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431"} err="failed to get container status \"f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431\": rpc error: code = NotFound desc = could not find container \"f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431\": container with ID starting with f620183e84a7a018a6ba9fe5db93849d8979e69e362a1b92d496d0a3a8f33431 not found: ID does not exist" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.285596 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm"] Dec 03 17:18:10 crc kubenswrapper[4787]: E1203 17:18:10.285835 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="670e0167-14c0-4547-9539-f303e18c33f3" containerName="route-controller-manager" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.285847 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="670e0167-14c0-4547-9539-f303e18c33f3" containerName="route-controller-manager" Dec 03 17:18:10 crc kubenswrapper[4787]: E1203 17:18:10.285861 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f41a2379-8c92-4d95-96cc-3d466bdcbc1d" containerName="controller-manager" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.285867 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f41a2379-8c92-4d95-96cc-3d466bdcbc1d" containerName="controller-manager" Dec 03 17:18:10 crc kubenswrapper[4787]: E1203 17:18:10.285882 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.285888 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.285995 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.286008 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="670e0167-14c0-4547-9539-f303e18c33f3" containerName="route-controller-manager" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.286042 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f41a2379-8c92-4d95-96cc-3d466bdcbc1d" containerName="controller-manager" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.286583 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.290074 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p8twg"] Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.290465 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.290785 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.291203 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.291045 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.291324 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.291353 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.291150 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.296933 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm"] Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.301934 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.302376 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.302491 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.302750 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.302862 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.302964 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.305661 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.353874 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p8twg"] Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.383453 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df7fa071-55aa-4b52-b578-288896a42043-serving-cert\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.383499 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lm2fh\" (UniqueName: \"kubernetes.io/projected/9b9a913e-844c-44b9-b192-61b1dc0d65a6-kube-api-access-lm2fh\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.383529 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs8cq\" (UniqueName: \"kubernetes.io/projected/df7fa071-55aa-4b52-b578-288896a42043-kube-api-access-gs8cq\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.383597 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-config\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.383629 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-config\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.383661 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b9a913e-844c-44b9-b192-61b1dc0d65a6-serving-cert\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.383696 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-client-ca\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.383712 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-client-ca\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.383777 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-proxy-ca-bundles\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.485278 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs8cq\" (UniqueName: \"kubernetes.io/projected/df7fa071-55aa-4b52-b578-288896a42043-kube-api-access-gs8cq\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.485372 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-config\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.485428 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-config\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.485464 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b9a913e-844c-44b9-b192-61b1dc0d65a6-serving-cert\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.485509 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-client-ca\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.485540 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-client-ca\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.485599 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-proxy-ca-bundles\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.485673 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lm2fh\" (UniqueName: \"kubernetes.io/projected/9b9a913e-844c-44b9-b192-61b1dc0d65a6-kube-api-access-lm2fh\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.485707 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df7fa071-55aa-4b52-b578-288896a42043-serving-cert\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.487779 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-config\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.487801 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-client-ca\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.489547 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b9a913e-844c-44b9-b192-61b1dc0d65a6-serving-cert\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.489856 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-config\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.490271 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-client-ca\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.491418 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-proxy-ca-bundles\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.500097 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df7fa071-55aa-4b52-b578-288896a42043-serving-cert\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.515369 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lm2fh\" (UniqueName: \"kubernetes.io/projected/9b9a913e-844c-44b9-b192-61b1dc0d65a6-kube-api-access-lm2fh\") pod \"route-controller-manager-558cd6b58-lb9hm\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.517647 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs8cq\" (UniqueName: \"kubernetes.io/projected/df7fa071-55aa-4b52-b578-288896a42043-kube-api-access-gs8cq\") pod \"controller-manager-56b87fd594-p8twg\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.655232 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.663161 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.869472 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm"] Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.926095 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p8twg"] Dec 03 17:18:10 crc kubenswrapper[4787]: W1203 17:18:10.935838 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf7fa071_55aa_4b52_b578_288896a42043.slice/crio-bdd930a74fa66bb8130598af61ccb9b12745aaeafe5b2a6446648b52be1f758a WatchSource:0}: Error finding container bdd930a74fa66bb8130598af61ccb9b12745aaeafe5b2a6446648b52be1f758a: Status 404 returned error can't find the container with id bdd930a74fa66bb8130598af61ccb9b12745aaeafe5b2a6446648b52be1f758a Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.983377 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" event={"ID":"df7fa071-55aa-4b52-b578-288896a42043","Type":"ContainerStarted","Data":"bdd930a74fa66bb8130598af61ccb9b12745aaeafe5b2a6446648b52be1f758a"} Dec 03 17:18:10 crc kubenswrapper[4787]: I1203 17:18:10.985351 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" event={"ID":"9b9a913e-844c-44b9-b192-61b1dc0d65a6","Type":"ContainerStarted","Data":"5882e232f9ee39d12ccd2010420b00092dacf11df9d87339ebdc747e2b385e39"} Dec 03 17:18:11 crc kubenswrapper[4787]: I1203 17:18:11.777004 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="670e0167-14c0-4547-9539-f303e18c33f3" path="/var/lib/kubelet/pods/670e0167-14c0-4547-9539-f303e18c33f3/volumes" Dec 03 17:18:11 crc kubenswrapper[4787]: I1203 17:18:11.778172 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f41a2379-8c92-4d95-96cc-3d466bdcbc1d" path="/var/lib/kubelet/pods/f41a2379-8c92-4d95-96cc-3d466bdcbc1d/volumes" Dec 03 17:18:11 crc kubenswrapper[4787]: I1203 17:18:11.990521 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" event={"ID":"9b9a913e-844c-44b9-b192-61b1dc0d65a6","Type":"ContainerStarted","Data":"291f1e6e5ccb4242b892e1aa86cd6d33b762253d0a02e9bc1d032b0cb41ace16"} Dec 03 17:18:11 crc kubenswrapper[4787]: I1203 17:18:11.990609 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:11 crc kubenswrapper[4787]: I1203 17:18:11.993224 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" event={"ID":"df7fa071-55aa-4b52-b578-288896a42043","Type":"ContainerStarted","Data":"7f19c46ecc0ef0805c6dd44e00c76c1252e9a102dfa83e236932314c42ca143b"} Dec 03 17:18:11 crc kubenswrapper[4787]: I1203 17:18:11.993505 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:11 crc kubenswrapper[4787]: I1203 17:18:11.997511 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:11 crc kubenswrapper[4787]: I1203 17:18:11.998549 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:12 crc kubenswrapper[4787]: I1203 17:18:12.008079 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" podStartSLOduration=3.008006783 podStartE2EDuration="3.008006783s" podCreationTimestamp="2025-12-03 17:18:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:18:12.006672236 +0000 UTC m=+308.824143215" watchObservedRunningTime="2025-12-03 17:18:12.008006783 +0000 UTC m=+308.825477742" Dec 03 17:18:12 crc kubenswrapper[4787]: I1203 17:18:12.028534 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" podStartSLOduration=3.028517787 podStartE2EDuration="3.028517787s" podCreationTimestamp="2025-12-03 17:18:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:18:12.025373709 +0000 UTC m=+308.842844678" watchObservedRunningTime="2025-12-03 17:18:12.028517787 +0000 UTC m=+308.845988746" Dec 03 17:18:21 crc kubenswrapper[4787]: I1203 17:18:21.486953 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p8twg"] Dec 03 17:18:21 crc kubenswrapper[4787]: I1203 17:18:21.488332 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" podUID="df7fa071-55aa-4b52-b578-288896a42043" containerName="controller-manager" containerID="cri-o://7f19c46ecc0ef0805c6dd44e00c76c1252e9a102dfa83e236932314c42ca143b" gracePeriod=30 Dec 03 17:18:21 crc kubenswrapper[4787]: I1203 17:18:21.500700 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm"] Dec 03 17:18:21 crc kubenswrapper[4787]: I1203 17:18:21.501012 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" podUID="9b9a913e-844c-44b9-b192-61b1dc0d65a6" containerName="route-controller-manager" containerID="cri-o://291f1e6e5ccb4242b892e1aa86cd6d33b762253d0a02e9bc1d032b0cb41ace16" gracePeriod=30 Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.060419 4787 generic.go:334] "Generic (PLEG): container finished" podID="9b9a913e-844c-44b9-b192-61b1dc0d65a6" containerID="291f1e6e5ccb4242b892e1aa86cd6d33b762253d0a02e9bc1d032b0cb41ace16" exitCode=0 Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.060513 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" event={"ID":"9b9a913e-844c-44b9-b192-61b1dc0d65a6","Type":"ContainerDied","Data":"291f1e6e5ccb4242b892e1aa86cd6d33b762253d0a02e9bc1d032b0cb41ace16"} Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.062090 4787 generic.go:334] "Generic (PLEG): container finished" podID="df7fa071-55aa-4b52-b578-288896a42043" containerID="7f19c46ecc0ef0805c6dd44e00c76c1252e9a102dfa83e236932314c42ca143b" exitCode=0 Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.062135 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" event={"ID":"df7fa071-55aa-4b52-b578-288896a42043","Type":"ContainerDied","Data":"7f19c46ecc0ef0805c6dd44e00c76c1252e9a102dfa83e236932314c42ca143b"} Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.515201 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.576155 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lm2fh\" (UniqueName: \"kubernetes.io/projected/9b9a913e-844c-44b9-b192-61b1dc0d65a6-kube-api-access-lm2fh\") pod \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.576388 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-config\") pod \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.576557 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b9a913e-844c-44b9-b192-61b1dc0d65a6-serving-cert\") pod \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.576668 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-client-ca\") pod \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\" (UID: \"9b9a913e-844c-44b9-b192-61b1dc0d65a6\") " Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.577071 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-config" (OuterVolumeSpecName: "config") pod "9b9a913e-844c-44b9-b192-61b1dc0d65a6" (UID: "9b9a913e-844c-44b9-b192-61b1dc0d65a6"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.577269 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-client-ca" (OuterVolumeSpecName: "client-ca") pod "9b9a913e-844c-44b9-b192-61b1dc0d65a6" (UID: "9b9a913e-844c-44b9-b192-61b1dc0d65a6"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.582167 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b9a913e-844c-44b9-b192-61b1dc0d65a6-kube-api-access-lm2fh" (OuterVolumeSpecName: "kube-api-access-lm2fh") pod "9b9a913e-844c-44b9-b192-61b1dc0d65a6" (UID: "9b9a913e-844c-44b9-b192-61b1dc0d65a6"). InnerVolumeSpecName "kube-api-access-lm2fh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.582401 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9b9a913e-844c-44b9-b192-61b1dc0d65a6-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9b9a913e-844c-44b9-b192-61b1dc0d65a6" (UID: "9b9a913e-844c-44b9-b192-61b1dc0d65a6"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.678640 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lm2fh\" (UniqueName: \"kubernetes.io/projected/9b9a913e-844c-44b9-b192-61b1dc0d65a6-kube-api-access-lm2fh\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.678774 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.678799 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9b9a913e-844c-44b9-b192-61b1dc0d65a6-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.678813 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9b9a913e-844c-44b9-b192-61b1dc0d65a6-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.688993 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.779283 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-client-ca\") pod \"df7fa071-55aa-4b52-b578-288896a42043\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.779353 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-config\") pod \"df7fa071-55aa-4b52-b578-288896a42043\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.779456 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df7fa071-55aa-4b52-b578-288896a42043-serving-cert\") pod \"df7fa071-55aa-4b52-b578-288896a42043\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.779517 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-proxy-ca-bundles\") pod \"df7fa071-55aa-4b52-b578-288896a42043\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.779559 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs8cq\" (UniqueName: \"kubernetes.io/projected/df7fa071-55aa-4b52-b578-288896a42043-kube-api-access-gs8cq\") pod \"df7fa071-55aa-4b52-b578-288896a42043\" (UID: \"df7fa071-55aa-4b52-b578-288896a42043\") " Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.781043 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "df7fa071-55aa-4b52-b578-288896a42043" (UID: "df7fa071-55aa-4b52-b578-288896a42043"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.781073 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-client-ca" (OuterVolumeSpecName: "client-ca") pod "df7fa071-55aa-4b52-b578-288896a42043" (UID: "df7fa071-55aa-4b52-b578-288896a42043"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.781100 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-config" (OuterVolumeSpecName: "config") pod "df7fa071-55aa-4b52-b578-288896a42043" (UID: "df7fa071-55aa-4b52-b578-288896a42043"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.794099 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df7fa071-55aa-4b52-b578-288896a42043-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "df7fa071-55aa-4b52-b578-288896a42043" (UID: "df7fa071-55aa-4b52-b578-288896a42043"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.795061 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df7fa071-55aa-4b52-b578-288896a42043-kube-api-access-gs8cq" (OuterVolumeSpecName: "kube-api-access-gs8cq") pod "df7fa071-55aa-4b52-b578-288896a42043" (UID: "df7fa071-55aa-4b52-b578-288896a42043"). InnerVolumeSpecName "kube-api-access-gs8cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.881408 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.881479 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs8cq\" (UniqueName: \"kubernetes.io/projected/df7fa071-55aa-4b52-b578-288896a42043-kube-api-access-gs8cq\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.881489 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.881498 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df7fa071-55aa-4b52-b578-288896a42043-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:22 crc kubenswrapper[4787]: I1203 17:18:22.881506 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/df7fa071-55aa-4b52-b578-288896a42043-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.069330 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" event={"ID":"df7fa071-55aa-4b52-b578-288896a42043","Type":"ContainerDied","Data":"bdd930a74fa66bb8130598af61ccb9b12745aaeafe5b2a6446648b52be1f758a"} Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.069664 4787 scope.go:117] "RemoveContainer" containerID="7f19c46ecc0ef0805c6dd44e00c76c1252e9a102dfa83e236932314c42ca143b" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.069355 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-56b87fd594-p8twg" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.071281 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" event={"ID":"9b9a913e-844c-44b9-b192-61b1dc0d65a6","Type":"ContainerDied","Data":"5882e232f9ee39d12ccd2010420b00092dacf11df9d87339ebdc747e2b385e39"} Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.071330 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.088816 4787 scope.go:117] "RemoveContainer" containerID="291f1e6e5ccb4242b892e1aa86cd6d33b762253d0a02e9bc1d032b0cb41ace16" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.109198 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p8twg"] Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.120754 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-56b87fd594-p8twg"] Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.129560 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm"] Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.133580 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-558cd6b58-lb9hm"] Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.287123 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-4smbh"] Dec 03 17:18:23 crc kubenswrapper[4787]: E1203 17:18:23.287326 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df7fa071-55aa-4b52-b578-288896a42043" containerName="controller-manager" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.287338 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="df7fa071-55aa-4b52-b578-288896a42043" containerName="controller-manager" Dec 03 17:18:23 crc kubenswrapper[4787]: E1203 17:18:23.287354 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b9a913e-844c-44b9-b192-61b1dc0d65a6" containerName="route-controller-manager" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.287360 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b9a913e-844c-44b9-b192-61b1dc0d65a6" containerName="route-controller-manager" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.287455 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="df7fa071-55aa-4b52-b578-288896a42043" containerName="controller-manager" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.287467 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b9a913e-844c-44b9-b192-61b1dc0d65a6" containerName="route-controller-manager" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.287802 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.290146 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.290224 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.290394 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.290494 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.290760 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.293485 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49"] Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.295078 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.295356 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.299269 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.299493 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49"] Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.299570 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.299578 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.299652 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.299824 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.299859 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.302546 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.313818 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-4smbh"] Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.387587 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2gkh\" (UniqueName: \"kubernetes.io/projected/b15b9167-088c-44e9-b535-59abeeb65d47-kube-api-access-j2gkh\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.387652 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-client-ca\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.387723 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-config\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.387748 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-client-ca\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.387775 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-proxy-ca-bundles\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.387850 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d9p7\" (UniqueName: \"kubernetes.io/projected/a55d50b7-531c-4b44-8951-2ac7054fca09-kube-api-access-8d9p7\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.387886 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-config\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.387908 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a55d50b7-531c-4b44-8951-2ac7054fca09-serving-cert\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.387937 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b15b9167-088c-44e9-b535-59abeeb65d47-serving-cert\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.489501 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2gkh\" (UniqueName: \"kubernetes.io/projected/b15b9167-088c-44e9-b535-59abeeb65d47-kube-api-access-j2gkh\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.490166 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-client-ca\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.490277 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-config\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.490304 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-client-ca\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.490366 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-proxy-ca-bundles\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.490456 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d9p7\" (UniqueName: \"kubernetes.io/projected/a55d50b7-531c-4b44-8951-2ac7054fca09-kube-api-access-8d9p7\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.490488 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-config\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.490519 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a55d50b7-531c-4b44-8951-2ac7054fca09-serving-cert\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.490563 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b15b9167-088c-44e9-b535-59abeeb65d47-serving-cert\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.491695 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-proxy-ca-bundles\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.491884 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-client-ca\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.492057 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-config\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.492084 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-config\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.492293 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-client-ca\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.495169 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a55d50b7-531c-4b44-8951-2ac7054fca09-serving-cert\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.498668 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b15b9167-088c-44e9-b535-59abeeb65d47-serving-cert\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.505215 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2gkh\" (UniqueName: \"kubernetes.io/projected/b15b9167-088c-44e9-b535-59abeeb65d47-kube-api-access-j2gkh\") pod \"controller-manager-5967c84899-4smbh\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.506695 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d9p7\" (UniqueName: \"kubernetes.io/projected/a55d50b7-531c-4b44-8951-2ac7054fca09-kube-api-access-8d9p7\") pod \"route-controller-manager-bcfc6b96c-hhk49\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.605503 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.612600 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.779100 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b9a913e-844c-44b9-b192-61b1dc0d65a6" path="/var/lib/kubelet/pods/9b9a913e-844c-44b9-b192-61b1dc0d65a6/volumes" Dec 03 17:18:23 crc kubenswrapper[4787]: I1203 17:18:23.780626 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df7fa071-55aa-4b52-b578-288896a42043" path="/var/lib/kubelet/pods/df7fa071-55aa-4b52-b578-288896a42043/volumes" Dec 03 17:18:24 crc kubenswrapper[4787]: I1203 17:18:24.045804 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-4smbh"] Dec 03 17:18:24 crc kubenswrapper[4787]: W1203 17:18:24.053871 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb15b9167_088c_44e9_b535_59abeeb65d47.slice/crio-fcd70177751fece332128e3fb022147b0a49812fe66626e2c78aa2f65d47358f WatchSource:0}: Error finding container fcd70177751fece332128e3fb022147b0a49812fe66626e2c78aa2f65d47358f: Status 404 returned error can't find the container with id fcd70177751fece332128e3fb022147b0a49812fe66626e2c78aa2f65d47358f Dec 03 17:18:24 crc kubenswrapper[4787]: I1203 17:18:24.079911 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" event={"ID":"b15b9167-088c-44e9-b535-59abeeb65d47","Type":"ContainerStarted","Data":"fcd70177751fece332128e3fb022147b0a49812fe66626e2c78aa2f65d47358f"} Dec 03 17:18:24 crc kubenswrapper[4787]: I1203 17:18:24.101664 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49"] Dec 03 17:18:25 crc kubenswrapper[4787]: I1203 17:18:25.089172 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" event={"ID":"a55d50b7-531c-4b44-8951-2ac7054fca09","Type":"ContainerStarted","Data":"1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967"} Dec 03 17:18:25 crc kubenswrapper[4787]: I1203 17:18:25.089246 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" event={"ID":"a55d50b7-531c-4b44-8951-2ac7054fca09","Type":"ContainerStarted","Data":"5114142910ac04bfe26112ee30563becac44250bbf5865fe761820814f1453c6"} Dec 03 17:18:25 crc kubenswrapper[4787]: I1203 17:18:25.089273 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:25 crc kubenswrapper[4787]: I1203 17:18:25.090800 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" event={"ID":"b15b9167-088c-44e9-b535-59abeeb65d47","Type":"ContainerStarted","Data":"d0f4fc5e2cc2892b2c2723620a655cc7cd78f90546504fd1e80dc18803d0f7d9"} Dec 03 17:18:25 crc kubenswrapper[4787]: I1203 17:18:25.091051 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:25 crc kubenswrapper[4787]: I1203 17:18:25.094888 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:18:25 crc kubenswrapper[4787]: I1203 17:18:25.098602 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:25 crc kubenswrapper[4787]: I1203 17:18:25.105363 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" podStartSLOduration=4.105338746 podStartE2EDuration="4.105338746s" podCreationTimestamp="2025-12-03 17:18:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:18:25.104159293 +0000 UTC m=+321.921630292" watchObservedRunningTime="2025-12-03 17:18:25.105338746 +0000 UTC m=+321.922809705" Dec 03 17:18:48 crc kubenswrapper[4787]: I1203 17:18:48.993145 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" podStartSLOduration=27.993122333 podStartE2EDuration="27.993122333s" podCreationTimestamp="2025-12-03 17:18:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:18:25.146964221 +0000 UTC m=+321.964435180" watchObservedRunningTime="2025-12-03 17:18:48.993122333 +0000 UTC m=+345.810593332" Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.000754 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-4smbh"] Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.001182 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" podUID="b15b9167-088c-44e9-b535-59abeeb65d47" containerName="controller-manager" containerID="cri-o://d0f4fc5e2cc2892b2c2723620a655cc7cd78f90546504fd1e80dc18803d0f7d9" gracePeriod=30 Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.236665 4787 generic.go:334] "Generic (PLEG): container finished" podID="b15b9167-088c-44e9-b535-59abeeb65d47" containerID="d0f4fc5e2cc2892b2c2723620a655cc7cd78f90546504fd1e80dc18803d0f7d9" exitCode=0 Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.236714 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" event={"ID":"b15b9167-088c-44e9-b535-59abeeb65d47","Type":"ContainerDied","Data":"d0f4fc5e2cc2892b2c2723620a655cc7cd78f90546504fd1e80dc18803d0f7d9"} Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.609277 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.747384 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-proxy-ca-bundles\") pod \"b15b9167-088c-44e9-b535-59abeeb65d47\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.747430 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b15b9167-088c-44e9-b535-59abeeb65d47-serving-cert\") pod \"b15b9167-088c-44e9-b535-59abeeb65d47\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.747476 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2gkh\" (UniqueName: \"kubernetes.io/projected/b15b9167-088c-44e9-b535-59abeeb65d47-kube-api-access-j2gkh\") pod \"b15b9167-088c-44e9-b535-59abeeb65d47\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.747541 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-client-ca\") pod \"b15b9167-088c-44e9-b535-59abeeb65d47\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.747621 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-config\") pod \"b15b9167-088c-44e9-b535-59abeeb65d47\" (UID: \"b15b9167-088c-44e9-b535-59abeeb65d47\") " Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.748683 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b15b9167-088c-44e9-b535-59abeeb65d47" (UID: "b15b9167-088c-44e9-b535-59abeeb65d47"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.748697 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-config" (OuterVolumeSpecName: "config") pod "b15b9167-088c-44e9-b535-59abeeb65d47" (UID: "b15b9167-088c-44e9-b535-59abeeb65d47"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.748980 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-client-ca" (OuterVolumeSpecName: "client-ca") pod "b15b9167-088c-44e9-b535-59abeeb65d47" (UID: "b15b9167-088c-44e9-b535-59abeeb65d47"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.753410 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b15b9167-088c-44e9-b535-59abeeb65d47-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b15b9167-088c-44e9-b535-59abeeb65d47" (UID: "b15b9167-088c-44e9-b535-59abeeb65d47"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.755351 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b15b9167-088c-44e9-b535-59abeeb65d47-kube-api-access-j2gkh" (OuterVolumeSpecName: "kube-api-access-j2gkh") pod "b15b9167-088c-44e9-b535-59abeeb65d47" (UID: "b15b9167-088c-44e9-b535-59abeeb65d47"). InnerVolumeSpecName "kube-api-access-j2gkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.849493 4787 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.849551 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b15b9167-088c-44e9-b535-59abeeb65d47-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.849564 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2gkh\" (UniqueName: \"kubernetes.io/projected/b15b9167-088c-44e9-b535-59abeeb65d47-kube-api-access-j2gkh\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.849578 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:49 crc kubenswrapper[4787]: I1203 17:18:49.849591 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b15b9167-088c-44e9-b535-59abeeb65d47-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.245745 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" event={"ID":"b15b9167-088c-44e9-b535-59abeeb65d47","Type":"ContainerDied","Data":"fcd70177751fece332128e3fb022147b0a49812fe66626e2c78aa2f65d47358f"} Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.246412 4787 scope.go:117] "RemoveContainer" containerID="d0f4fc5e2cc2892b2c2723620a655cc7cd78f90546504fd1e80dc18803d0f7d9" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.245831 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5967c84899-4smbh" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.275510 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-4smbh"] Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.282101 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5967c84899-4smbh"] Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.307575 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-85446c8c96-hlv99"] Dec 03 17:18:50 crc kubenswrapper[4787]: E1203 17:18:50.307908 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b15b9167-088c-44e9-b535-59abeeb65d47" containerName="controller-manager" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.307929 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b15b9167-088c-44e9-b535-59abeeb65d47" containerName="controller-manager" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.308076 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b15b9167-088c-44e9-b535-59abeeb65d47" containerName="controller-manager" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.308572 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.313139 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.313176 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.313484 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.313734 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.314649 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.315507 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.317339 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85446c8c96-hlv99"] Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.323011 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.455530 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6600608e-598b-4e14-973c-f7aae250ead0-client-ca\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.455618 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqs2p\" (UniqueName: \"kubernetes.io/projected/6600608e-598b-4e14-973c-f7aae250ead0-kube-api-access-qqs2p\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.455641 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6600608e-598b-4e14-973c-f7aae250ead0-serving-cert\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.455679 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6600608e-598b-4e14-973c-f7aae250ead0-proxy-ca-bundles\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.455706 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6600608e-598b-4e14-973c-f7aae250ead0-config\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.556391 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6600608e-598b-4e14-973c-f7aae250ead0-config\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.556759 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6600608e-598b-4e14-973c-f7aae250ead0-client-ca\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.556820 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqs2p\" (UniqueName: \"kubernetes.io/projected/6600608e-598b-4e14-973c-f7aae250ead0-kube-api-access-qqs2p\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.556847 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6600608e-598b-4e14-973c-f7aae250ead0-serving-cert\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.556900 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6600608e-598b-4e14-973c-f7aae250ead0-proxy-ca-bundles\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.558357 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6600608e-598b-4e14-973c-f7aae250ead0-proxy-ca-bundles\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.559211 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6600608e-598b-4e14-973c-f7aae250ead0-client-ca\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.559331 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6600608e-598b-4e14-973c-f7aae250ead0-config\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.562869 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6600608e-598b-4e14-973c-f7aae250ead0-serving-cert\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.586313 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqs2p\" (UniqueName: \"kubernetes.io/projected/6600608e-598b-4e14-973c-f7aae250ead0-kube-api-access-qqs2p\") pod \"controller-manager-85446c8c96-hlv99\" (UID: \"6600608e-598b-4e14-973c-f7aae250ead0\") " pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:50 crc kubenswrapper[4787]: I1203 17:18:50.623822 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:51 crc kubenswrapper[4787]: I1203 17:18:51.057473 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-85446c8c96-hlv99"] Dec 03 17:18:51 crc kubenswrapper[4787]: I1203 17:18:51.251853 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" event={"ID":"6600608e-598b-4e14-973c-f7aae250ead0","Type":"ContainerStarted","Data":"b055d556ff82cc8a2a1a6ddaab3dbc7eee4d2921bfb42cc6137dce19836e0eb8"} Dec 03 17:18:51 crc kubenswrapper[4787]: I1203 17:18:51.251922 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" event={"ID":"6600608e-598b-4e14-973c-f7aae250ead0","Type":"ContainerStarted","Data":"2c9ea7b3ce56aa8048d86f64e7abbf9b085d1b67ced2ae6b8b4818811b22eb11"} Dec 03 17:18:51 crc kubenswrapper[4787]: I1203 17:18:51.252205 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:51 crc kubenswrapper[4787]: I1203 17:18:51.253895 4787 patch_prober.go:28] interesting pod/controller-manager-85446c8c96-hlv99 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" start-of-body= Dec 03 17:18:51 crc kubenswrapper[4787]: I1203 17:18:51.253948 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" podUID="6600608e-598b-4e14-973c-f7aae250ead0" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.62:8443/healthz\": dial tcp 10.217.0.62:8443: connect: connection refused" Dec 03 17:18:51 crc kubenswrapper[4787]: I1203 17:18:51.771783 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b15b9167-088c-44e9-b535-59abeeb65d47" path="/var/lib/kubelet/pods/b15b9167-088c-44e9-b535-59abeeb65d47/volumes" Dec 03 17:18:52 crc kubenswrapper[4787]: I1203 17:18:52.263080 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" Dec 03 17:18:52 crc kubenswrapper[4787]: I1203 17:18:52.280943 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-85446c8c96-hlv99" podStartSLOduration=3.28092279 podStartE2EDuration="3.28092279s" podCreationTimestamp="2025-12-03 17:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:18:51.269714213 +0000 UTC m=+348.087185172" watchObservedRunningTime="2025-12-03 17:18:52.28092279 +0000 UTC m=+349.098393759" Dec 03 17:19:08 crc kubenswrapper[4787]: I1203 17:19:08.978828 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49"] Dec 03 17:19:08 crc kubenswrapper[4787]: I1203 17:19:08.979679 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" podUID="a55d50b7-531c-4b44-8951-2ac7054fca09" containerName="route-controller-manager" containerID="cri-o://1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967" gracePeriod=30 Dec 03 17:19:09 crc kubenswrapper[4787]: I1203 17:19:09.925652 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.121187 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-client-ca\") pod \"a55d50b7-531c-4b44-8951-2ac7054fca09\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.121306 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8d9p7\" (UniqueName: \"kubernetes.io/projected/a55d50b7-531c-4b44-8951-2ac7054fca09-kube-api-access-8d9p7\") pod \"a55d50b7-531c-4b44-8951-2ac7054fca09\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.121370 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a55d50b7-531c-4b44-8951-2ac7054fca09-serving-cert\") pod \"a55d50b7-531c-4b44-8951-2ac7054fca09\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.121411 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-config\") pod \"a55d50b7-531c-4b44-8951-2ac7054fca09\" (UID: \"a55d50b7-531c-4b44-8951-2ac7054fca09\") " Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.122235 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-client-ca" (OuterVolumeSpecName: "client-ca") pod "a55d50b7-531c-4b44-8951-2ac7054fca09" (UID: "a55d50b7-531c-4b44-8951-2ac7054fca09"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.123062 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-config" (OuterVolumeSpecName: "config") pod "a55d50b7-531c-4b44-8951-2ac7054fca09" (UID: "a55d50b7-531c-4b44-8951-2ac7054fca09"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.131387 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a55d50b7-531c-4b44-8951-2ac7054fca09-kube-api-access-8d9p7" (OuterVolumeSpecName: "kube-api-access-8d9p7") pod "a55d50b7-531c-4b44-8951-2ac7054fca09" (UID: "a55d50b7-531c-4b44-8951-2ac7054fca09"). InnerVolumeSpecName "kube-api-access-8d9p7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.133351 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a55d50b7-531c-4b44-8951-2ac7054fca09-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "a55d50b7-531c-4b44-8951-2ac7054fca09" (UID: "a55d50b7-531c-4b44-8951-2ac7054fca09"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.223156 4787 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-client-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.223200 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8d9p7\" (UniqueName: \"kubernetes.io/projected/a55d50b7-531c-4b44-8951-2ac7054fca09-kube-api-access-8d9p7\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.223217 4787 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a55d50b7-531c-4b44-8951-2ac7054fca09-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.223228 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a55d50b7-531c-4b44-8951-2ac7054fca09-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.321932 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd"] Dec 03 17:19:10 crc kubenswrapper[4787]: E1203 17:19:10.322309 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a55d50b7-531c-4b44-8951-2ac7054fca09" containerName="route-controller-manager" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.322331 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a55d50b7-531c-4b44-8951-2ac7054fca09" containerName="route-controller-manager" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.322472 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a55d50b7-531c-4b44-8951-2ac7054fca09" containerName="route-controller-manager" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.323214 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.335476 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd"] Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.371098 4787 generic.go:334] "Generic (PLEG): container finished" podID="a55d50b7-531c-4b44-8951-2ac7054fca09" containerID="1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967" exitCode=0 Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.371145 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" event={"ID":"a55d50b7-531c-4b44-8951-2ac7054fca09","Type":"ContainerDied","Data":"1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967"} Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.371239 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" event={"ID":"a55d50b7-531c-4b44-8951-2ac7054fca09","Type":"ContainerDied","Data":"5114142910ac04bfe26112ee30563becac44250bbf5865fe761820814f1453c6"} Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.371286 4787 scope.go:117] "RemoveContainer" containerID="1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.371646 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.406210 4787 scope.go:117] "RemoveContainer" containerID="1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967" Dec 03 17:19:10 crc kubenswrapper[4787]: E1203 17:19:10.410204 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967\": container with ID starting with 1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967 not found: ID does not exist" containerID="1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.410296 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967"} err="failed to get container status \"1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967\": rpc error: code = NotFound desc = could not find container \"1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967\": container with ID starting with 1f5b907255511cc8745a3b971abfe033809710f3fea496b51d309329786cb967 not found: ID does not exist" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.426468 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lk5ch\" (UniqueName: \"kubernetes.io/projected/3dd4e780-c18c-4864-8a56-3ca3b442b96a-kube-api-access-lk5ch\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.426534 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dd4e780-c18c-4864-8a56-3ca3b442b96a-serving-cert\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.426650 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dd4e780-c18c-4864-8a56-3ca3b442b96a-config\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.426729 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3dd4e780-c18c-4864-8a56-3ca3b442b96a-client-ca\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.430748 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49"] Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.436780 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-bcfc6b96c-hhk49"] Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.527784 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lk5ch\" (UniqueName: \"kubernetes.io/projected/3dd4e780-c18c-4864-8a56-3ca3b442b96a-kube-api-access-lk5ch\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.527847 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dd4e780-c18c-4864-8a56-3ca3b442b96a-serving-cert\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.527884 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dd4e780-c18c-4864-8a56-3ca3b442b96a-config\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.527920 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3dd4e780-c18c-4864-8a56-3ca3b442b96a-client-ca\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.529892 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/3dd4e780-c18c-4864-8a56-3ca3b442b96a-client-ca\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.529974 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3dd4e780-c18c-4864-8a56-3ca3b442b96a-config\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.533228 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3dd4e780-c18c-4864-8a56-3ca3b442b96a-serving-cert\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.546591 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lk5ch\" (UniqueName: \"kubernetes.io/projected/3dd4e780-c18c-4864-8a56-3ca3b442b96a-kube-api-access-lk5ch\") pod \"route-controller-manager-69769948f6-kswhd\" (UID: \"3dd4e780-c18c-4864-8a56-3ca3b442b96a\") " pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:10 crc kubenswrapper[4787]: I1203 17:19:10.659259 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:11 crc kubenswrapper[4787]: I1203 17:19:11.052183 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd"] Dec 03 17:19:11 crc kubenswrapper[4787]: I1203 17:19:11.379007 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" event={"ID":"3dd4e780-c18c-4864-8a56-3ca3b442b96a","Type":"ContainerStarted","Data":"0febc4a433a07d3d86099c138af2d17545bf9db1392fc6e17408f36d014c8ed6"} Dec 03 17:19:11 crc kubenswrapper[4787]: I1203 17:19:11.380313 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" event={"ID":"3dd4e780-c18c-4864-8a56-3ca3b442b96a","Type":"ContainerStarted","Data":"031b60f95aba3d33b31b92b6283a4dd4fa1bc3acee1855bf4a43222de56753da"} Dec 03 17:19:11 crc kubenswrapper[4787]: I1203 17:19:11.380409 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:11 crc kubenswrapper[4787]: I1203 17:19:11.407897 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" podStartSLOduration=3.407865034 podStartE2EDuration="3.407865034s" podCreationTimestamp="2025-12-03 17:19:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:11.399470467 +0000 UTC m=+368.216941446" watchObservedRunningTime="2025-12-03 17:19:11.407865034 +0000 UTC m=+368.225336033" Dec 03 17:19:11 crc kubenswrapper[4787]: I1203 17:19:11.572162 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-69769948f6-kswhd" Dec 03 17:19:11 crc kubenswrapper[4787]: I1203 17:19:11.778366 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a55d50b7-531c-4b44-8951-2ac7054fca09" path="/var/lib/kubelet/pods/a55d50b7-531c-4b44-8951-2ac7054fca09/volumes" Dec 03 17:19:18 crc kubenswrapper[4787]: I1203 17:19:18.990230 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:19:18 crc kubenswrapper[4787]: I1203 17:19:18.990804 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.176037 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vxhgk"] Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.176351 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vxhgk" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" containerName="registry-server" containerID="cri-o://31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30" gracePeriod=30 Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.187033 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7zm2"] Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.187316 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-t7zm2" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerName="registry-server" containerID="cri-o://e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b" gracePeriod=30 Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.200490 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bl7vk"] Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.200761 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" podUID="684133b3-732f-47c6-b15d-108c3e37397f" containerName="marketplace-operator" containerID="cri-o://83490f7b9a6f69c61f85d85df8bb788c0c19df6c6daf4cc3344c2204d85b4497" gracePeriod=30 Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.221132 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-54jtg"] Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.221444 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-54jtg" podUID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" containerName="registry-server" containerID="cri-o://d49d37b6a018ec45695946ec36a92ba03a7ff7a992081fa39a04352a2f3ea43b" gracePeriod=30 Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.225704 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-278bm"] Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.226527 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.233239 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vnqdl"] Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.234396 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vnqdl" podUID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" containerName="registry-server" containerID="cri-o://4381fe49861a72228d6c3e8f60f3fbd18b8c01d4903b8becb40228b92d50a150" gracePeriod=30 Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.245558 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-278bm"] Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.260910 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zk7g8\" (UniqueName: \"kubernetes.io/projected/38b06bc8-6c75-46f8-a53b-d95109183306-kube-api-access-zk7g8\") pod \"marketplace-operator-79b997595-278bm\" (UID: \"38b06bc8-6c75-46f8-a53b-d95109183306\") " pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.260952 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/38b06bc8-6c75-46f8-a53b-d95109183306-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-278bm\" (UID: \"38b06bc8-6c75-46f8-a53b-d95109183306\") " pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.260982 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/38b06bc8-6c75-46f8-a53b-d95109183306-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-278bm\" (UID: \"38b06bc8-6c75-46f8-a53b-d95109183306\") " pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.363081 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zk7g8\" (UniqueName: \"kubernetes.io/projected/38b06bc8-6c75-46f8-a53b-d95109183306-kube-api-access-zk7g8\") pod \"marketplace-operator-79b997595-278bm\" (UID: \"38b06bc8-6c75-46f8-a53b-d95109183306\") " pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.363155 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/38b06bc8-6c75-46f8-a53b-d95109183306-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-278bm\" (UID: \"38b06bc8-6c75-46f8-a53b-d95109183306\") " pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.363194 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/38b06bc8-6c75-46f8-a53b-d95109183306-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-278bm\" (UID: \"38b06bc8-6c75-46f8-a53b-d95109183306\") " pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.364948 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/38b06bc8-6c75-46f8-a53b-d95109183306-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-278bm\" (UID: \"38b06bc8-6c75-46f8-a53b-d95109183306\") " pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.378319 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/38b06bc8-6c75-46f8-a53b-d95109183306-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-278bm\" (UID: \"38b06bc8-6c75-46f8-a53b-d95109183306\") " pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.382721 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zk7g8\" (UniqueName: \"kubernetes.io/projected/38b06bc8-6c75-46f8-a53b-d95109183306-kube-api-access-zk7g8\") pod \"marketplace-operator-79b997595-278bm\" (UID: \"38b06bc8-6c75-46f8-a53b-d95109183306\") " pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.440339 4787 generic.go:334] "Generic (PLEG): container finished" podID="684133b3-732f-47c6-b15d-108c3e37397f" containerID="83490f7b9a6f69c61f85d85df8bb788c0c19df6c6daf4cc3344c2204d85b4497" exitCode=0 Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.440420 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" event={"ID":"684133b3-732f-47c6-b15d-108c3e37397f","Type":"ContainerDied","Data":"83490f7b9a6f69c61f85d85df8bb788c0c19df6c6daf4cc3344c2204d85b4497"} Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.440461 4787 scope.go:117] "RemoveContainer" containerID="9cbe8ac34d758da238115f28c4c930ec9270479e51b277a9c2d76cba1e2ef25a" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.443768 4787 generic.go:334] "Generic (PLEG): container finished" podID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" containerID="d49d37b6a018ec45695946ec36a92ba03a7ff7a992081fa39a04352a2f3ea43b" exitCode=0 Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.443814 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54jtg" event={"ID":"2b45ac06-f7eb-47b8-b1c8-af110a7da844","Type":"ContainerDied","Data":"d49d37b6a018ec45695946ec36a92ba03a7ff7a992081fa39a04352a2f3ea43b"} Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.446463 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce9df171-a517-4444-9c60-589f83e29e8a" containerID="31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30" exitCode=0 Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.446511 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxhgk" event={"ID":"ce9df171-a517-4444-9c60-589f83e29e8a","Type":"ContainerDied","Data":"31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30"} Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.450798 4787 generic.go:334] "Generic (PLEG): container finished" podID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" containerID="4381fe49861a72228d6c3e8f60f3fbd18b8c01d4903b8becb40228b92d50a150" exitCode=0 Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.450867 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnqdl" event={"ID":"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43","Type":"ContainerDied","Data":"4381fe49861a72228d6c3e8f60f3fbd18b8c01d4903b8becb40228b92d50a150"} Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.454435 4787 generic.go:334] "Generic (PLEG): container finished" podID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerID="e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b" exitCode=0 Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.454471 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7zm2" event={"ID":"b4b8d0ee-3c0b-4a5b-9040-b650af634353","Type":"ContainerDied","Data":"e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b"} Dec 03 17:19:19 crc kubenswrapper[4787]: E1203 17:19:19.457642 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b is running failed: container process not found" containerID="e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 17:19:19 crc kubenswrapper[4787]: E1203 17:19:19.457944 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b is running failed: container process not found" containerID="e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 17:19:19 crc kubenswrapper[4787]: E1203 17:19:19.458243 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b is running failed: container process not found" containerID="e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 17:19:19 crc kubenswrapper[4787]: E1203 17:19:19.458283 4787 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/community-operators-t7zm2" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerName="registry-server" Dec 03 17:19:19 crc kubenswrapper[4787]: E1203 17:19:19.673209 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30 is running failed: container process not found" containerID="31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 17:19:19 crc kubenswrapper[4787]: E1203 17:19:19.677142 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30 is running failed: container process not found" containerID="31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 17:19:19 crc kubenswrapper[4787]: E1203 17:19:19.678246 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30 is running failed: container process not found" containerID="31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30" cmd=["grpc_health_probe","-addr=:50051"] Dec 03 17:19:19 crc kubenswrapper[4787]: E1203 17:19:19.678284 4787 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/certified-operators-vxhgk" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" containerName="registry-server" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.688184 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.744726 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.774922 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-trusted-ca\") pod \"684133b3-732f-47c6-b15d-108c3e37397f\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.774978 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb5lp\" (UniqueName: \"kubernetes.io/projected/684133b3-732f-47c6-b15d-108c3e37397f-kube-api-access-mb5lp\") pod \"684133b3-732f-47c6-b15d-108c3e37397f\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.775003 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-operator-metrics\") pod \"684133b3-732f-47c6-b15d-108c3e37397f\" (UID: \"684133b3-732f-47c6-b15d-108c3e37397f\") " Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.777863 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "684133b3-732f-47c6-b15d-108c3e37397f" (UID: "684133b3-732f-47c6-b15d-108c3e37397f"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.789476 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "684133b3-732f-47c6-b15d-108c3e37397f" (UID: "684133b3-732f-47c6-b15d-108c3e37397f"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.789673 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/684133b3-732f-47c6-b15d-108c3e37397f-kube-api-access-mb5lp" (OuterVolumeSpecName: "kube-api-access-mb5lp") pod "684133b3-732f-47c6-b15d-108c3e37397f" (UID: "684133b3-732f-47c6-b15d-108c3e37397f"). InnerVolumeSpecName "kube-api-access-mb5lp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.879705 4787 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.879748 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb5lp\" (UniqueName: \"kubernetes.io/projected/684133b3-732f-47c6-b15d-108c3e37397f-kube-api-access-mb5lp\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.879765 4787 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/684133b3-732f-47c6-b15d-108c3e37397f-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.937974 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.966173 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.969215 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.980612 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-utilities\") pod \"ce9df171-a517-4444-9c60-589f83e29e8a\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.980779 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmtr8\" (UniqueName: \"kubernetes.io/projected/ce9df171-a517-4444-9c60-589f83e29e8a-kube-api-access-tmtr8\") pod \"ce9df171-a517-4444-9c60-589f83e29e8a\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.980855 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-catalog-content\") pod \"ce9df171-a517-4444-9c60-589f83e29e8a\" (UID: \"ce9df171-a517-4444-9c60-589f83e29e8a\") " Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.982202 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-utilities" (OuterVolumeSpecName: "utilities") pod "ce9df171-a517-4444-9c60-589f83e29e8a" (UID: "ce9df171-a517-4444-9c60-589f83e29e8a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:19 crc kubenswrapper[4787]: I1203 17:19:19.988537 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce9df171-a517-4444-9c60-589f83e29e8a-kube-api-access-tmtr8" (OuterVolumeSpecName: "kube-api-access-tmtr8") pod "ce9df171-a517-4444-9c60-589f83e29e8a" (UID: "ce9df171-a517-4444-9c60-589f83e29e8a"). InnerVolumeSpecName "kube-api-access-tmtr8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.047229 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ce9df171-a517-4444-9c60-589f83e29e8a" (UID: "ce9df171-a517-4444-9c60-589f83e29e8a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.081865 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7876\" (UniqueName: \"kubernetes.io/projected/2b45ac06-f7eb-47b8-b1c8-af110a7da844-kube-api-access-w7876\") pod \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.081918 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-catalog-content\") pod \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.081943 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bbx4z\" (UniqueName: \"kubernetes.io/projected/b4b8d0ee-3c0b-4a5b-9040-b650af634353-kube-api-access-bbx4z\") pod \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.081988 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-catalog-content\") pod \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.082158 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-utilities\") pod \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\" (UID: \"2b45ac06-f7eb-47b8-b1c8-af110a7da844\") " Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.082198 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-utilities\") pod \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\" (UID: \"b4b8d0ee-3c0b-4a5b-9040-b650af634353\") " Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.082598 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.082617 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ce9df171-a517-4444-9c60-589f83e29e8a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.082629 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmtr8\" (UniqueName: \"kubernetes.io/projected/ce9df171-a517-4444-9c60-589f83e29e8a-kube-api-access-tmtr8\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.082981 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-utilities" (OuterVolumeSpecName: "utilities") pod "2b45ac06-f7eb-47b8-b1c8-af110a7da844" (UID: "2b45ac06-f7eb-47b8-b1c8-af110a7da844"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.083494 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-utilities" (OuterVolumeSpecName: "utilities") pod "b4b8d0ee-3c0b-4a5b-9040-b650af634353" (UID: "b4b8d0ee-3c0b-4a5b-9040-b650af634353"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.084679 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4b8d0ee-3c0b-4a5b-9040-b650af634353-kube-api-access-bbx4z" (OuterVolumeSpecName: "kube-api-access-bbx4z") pod "b4b8d0ee-3c0b-4a5b-9040-b650af634353" (UID: "b4b8d0ee-3c0b-4a5b-9040-b650af634353"). InnerVolumeSpecName "kube-api-access-bbx4z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.085795 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b45ac06-f7eb-47b8-b1c8-af110a7da844-kube-api-access-w7876" (OuterVolumeSpecName: "kube-api-access-w7876") pod "2b45ac06-f7eb-47b8-b1c8-af110a7da844" (UID: "2b45ac06-f7eb-47b8-b1c8-af110a7da844"). InnerVolumeSpecName "kube-api-access-w7876". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.086826 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.108474 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2b45ac06-f7eb-47b8-b1c8-af110a7da844" (UID: "2b45ac06-f7eb-47b8-b1c8-af110a7da844"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.144972 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4b8d0ee-3c0b-4a5b-9040-b650af634353" (UID: "b4b8d0ee-3c0b-4a5b-9040-b650af634353"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.183535 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7v49w\" (UniqueName: \"kubernetes.io/projected/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-kube-api-access-7v49w\") pod \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.183940 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-utilities\") pod \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.183981 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-catalog-content\") pod \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\" (UID: \"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43\") " Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.184310 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.184330 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.184339 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7876\" (UniqueName: \"kubernetes.io/projected/2b45ac06-f7eb-47b8-b1c8-af110a7da844-kube-api-access-w7876\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.184352 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b8d0ee-3c0b-4a5b-9040-b650af634353-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.184362 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bbx4z\" (UniqueName: \"kubernetes.io/projected/b4b8d0ee-3c0b-4a5b-9040-b650af634353-kube-api-access-bbx4z\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.184370 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2b45ac06-f7eb-47b8-b1c8-af110a7da844-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.184633 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-utilities" (OuterVolumeSpecName: "utilities") pod "b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" (UID: "b25eeef8-0ef5-44b4-acc5-d42ca33b3e43"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.186935 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-kube-api-access-7v49w" (OuterVolumeSpecName: "kube-api-access-7v49w") pod "b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" (UID: "b25eeef8-0ef5-44b4-acc5-d42ca33b3e43"). InnerVolumeSpecName "kube-api-access-7v49w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.242445 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-278bm"] Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.286133 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.286279 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7v49w\" (UniqueName: \"kubernetes.io/projected/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-kube-api-access-7v49w\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.301269 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" (UID: "b25eeef8-0ef5-44b4-acc5-d42ca33b3e43"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.387791 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.460675 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vnqdl" event={"ID":"b25eeef8-0ef5-44b4-acc5-d42ca33b3e43","Type":"ContainerDied","Data":"a9b11c2ff74affbfc48a0c8ed55a78472b53f763233cb6cddd93bfedefa53070"} Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.460712 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vnqdl" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.460724 4787 scope.go:117] "RemoveContainer" containerID="4381fe49861a72228d6c3e8f60f3fbd18b8c01d4903b8becb40228b92d50a150" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.462434 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-278bm" event={"ID":"38b06bc8-6c75-46f8-a53b-d95109183306","Type":"ContainerStarted","Data":"b6029f25bca7c21813c55118467a51024af5ebba2dbbc1cb5a29b73a50116f50"} Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.462486 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-278bm" event={"ID":"38b06bc8-6c75-46f8-a53b-d95109183306","Type":"ContainerStarted","Data":"d14dbacb0f8db5c3d5c9d024c527d2d1088431e626cc855d5bd99da325cae59c"} Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.462918 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.464246 4787 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-278bm container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.64:8080/healthz\": dial tcp 10.217.0.64:8080: connect: connection refused" start-of-body= Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.464315 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-278bm" podUID="38b06bc8-6c75-46f8-a53b-d95109183306" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.64:8080/healthz\": dial tcp 10.217.0.64:8080: connect: connection refused" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.466512 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-t7zm2" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.466509 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-t7zm2" event={"ID":"b4b8d0ee-3c0b-4a5b-9040-b650af634353","Type":"ContainerDied","Data":"56c7316818e625a5a976cc6c071fff16f56ee2f7953567b527f1d89b597ce4bc"} Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.468172 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.468187 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bl7vk" event={"ID":"684133b3-732f-47c6-b15d-108c3e37397f","Type":"ContainerDied","Data":"dd3a47df7001bc1a54a3c7851801c012a5d79a8cb331912d4e841556d4110921"} Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.472416 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-54jtg" event={"ID":"2b45ac06-f7eb-47b8-b1c8-af110a7da844","Type":"ContainerDied","Data":"09ba645de41ab67beacfd847a12a921b6497c40200ff16523b1071473494f4be"} Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.472435 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-54jtg" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.478218 4787 scope.go:117] "RemoveContainer" containerID="e0478542f27129f9f90f5aea14314e99e262284d82f8da7aa406f3977b7e8278" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.479969 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vxhgk" event={"ID":"ce9df171-a517-4444-9c60-589f83e29e8a","Type":"ContainerDied","Data":"0f880170fca6957d47634a0ab741c099bc0bb2711ef28cdc94ec81bee4ea6cae"} Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.480024 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vxhgk" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.496791 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-278bm" podStartSLOduration=1.496774901 podStartE2EDuration="1.496774901s" podCreationTimestamp="2025-12-03 17:19:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:20.490246236 +0000 UTC m=+377.307717195" watchObservedRunningTime="2025-12-03 17:19:20.496774901 +0000 UTC m=+377.314245870" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.513388 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vnqdl"] Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.514163 4787 scope.go:117] "RemoveContainer" containerID="40fe6a8a3887a1007d9f00a5e9d7a3c594b0d59972404137f599289c6d42ba05" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.521114 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vnqdl"] Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.530856 4787 scope.go:117] "RemoveContainer" containerID="e7d2b8235d7c8b5d82e6d574bf02740a639746a4918e015af0027c5f527e218b" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.533276 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-t7zm2"] Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.541277 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-t7zm2"] Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.544784 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bl7vk"] Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.552078 4787 scope.go:117] "RemoveContainer" containerID="5a5e127cd3ad536e0809114103e75e6b03b7bbb4ecdb35da2365d4cff2dd2849" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.564096 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bl7vk"] Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.585913 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vxhgk"] Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.589421 4787 scope.go:117] "RemoveContainer" containerID="a7bc0ada5e756e2efe42ba1498340a579e0518f9027865f66ddfc3277875a894" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.591227 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vxhgk"] Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.596328 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-54jtg"] Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.606558 4787 scope.go:117] "RemoveContainer" containerID="83490f7b9a6f69c61f85d85df8bb788c0c19df6c6daf4cc3344c2204d85b4497" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.607965 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-54jtg"] Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.623925 4787 scope.go:117] "RemoveContainer" containerID="d49d37b6a018ec45695946ec36a92ba03a7ff7a992081fa39a04352a2f3ea43b" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.638395 4787 scope.go:117] "RemoveContainer" containerID="507d5d030696c7bf38ef96a952b0b59248fffe7ae3780565c489d2a12e8cfc9d" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.653985 4787 scope.go:117] "RemoveContainer" containerID="74383b00206e80833ae9fb258dc9acce69fa06a1080fb3be2b3d65ed437f2972" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.674248 4787 scope.go:117] "RemoveContainer" containerID="31d3a9128c250dd623778fdc47a187ef7bcdee5375a9f2db8f86818688375a30" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.691248 4787 scope.go:117] "RemoveContainer" containerID="0108690ad93f843c0a2d787ae052bc7127ddfe527e17b6301cf693aa64dad5dc" Dec 03 17:19:20 crc kubenswrapper[4787]: I1203 17:19:20.707056 4787 scope.go:117] "RemoveContainer" containerID="2d46071bc3764a213bcfb154ec80446b7313243f16e78da67fb0e18d21c29ab6" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.495702 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-278bm" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.770714 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" path="/var/lib/kubelet/pods/2b45ac06-f7eb-47b8-b1c8-af110a7da844/volumes" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.771360 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="684133b3-732f-47c6-b15d-108c3e37397f" path="/var/lib/kubelet/pods/684133b3-732f-47c6-b15d-108c3e37397f/volumes" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.772764 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" path="/var/lib/kubelet/pods/b25eeef8-0ef5-44b4-acc5-d42ca33b3e43/volumes" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.773564 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" path="/var/lib/kubelet/pods/b4b8d0ee-3c0b-4a5b-9040-b650af634353/volumes" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.777982 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" path="/var/lib/kubelet/pods/ce9df171-a517-4444-9c60-589f83e29e8a/volumes" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.790611 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-rs8f5"] Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.790865 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" containerName="extract-utilities" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.790883 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" containerName="extract-utilities" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.790899 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="684133b3-732f-47c6-b15d-108c3e37397f" containerName="marketplace-operator" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.790907 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="684133b3-732f-47c6-b15d-108c3e37397f" containerName="marketplace-operator" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.790918 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" containerName="extract-content" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.790925 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" containerName="extract-content" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.790935 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.790942 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.790955 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.790965 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.790977 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerName="extract-utilities" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.790984 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerName="extract-utilities" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.790992 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" containerName="extract-content" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791000 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" containerName="extract-content" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.791008 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="684133b3-732f-47c6-b15d-108c3e37397f" containerName="marketplace-operator" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791038 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="684133b3-732f-47c6-b15d-108c3e37397f" containerName="marketplace-operator" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.791050 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791057 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.791068 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" containerName="extract-content" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791075 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" containerName="extract-content" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.791084 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" containerName="extract-utilities" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791093 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" containerName="extract-utilities" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.791103 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791110 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.791123 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" containerName="extract-utilities" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791131 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" containerName="extract-utilities" Dec 03 17:19:21 crc kubenswrapper[4787]: E1203 17:19:21.791145 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerName="extract-content" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791153 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerName="extract-content" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791241 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce9df171-a517-4444-9c60-589f83e29e8a" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791253 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="684133b3-732f-47c6-b15d-108c3e37397f" containerName="marketplace-operator" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791262 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b25eeef8-0ef5-44b4-acc5-d42ca33b3e43" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791278 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b45ac06-f7eb-47b8-b1c8-af110a7da844" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791288 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4b8d0ee-3c0b-4a5b-9040-b650af634353" containerName="registry-server" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791446 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="684133b3-732f-47c6-b15d-108c3e37397f" containerName="marketplace-operator" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.791951 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.794423 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.801707 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rs8f5"] Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.904962 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgchs\" (UniqueName: \"kubernetes.io/projected/1b309950-19b2-4d34-b4b8-f5336e6c47f5-kube-api-access-pgchs\") pod \"redhat-operators-rs8f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.905128 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-utilities\") pod \"redhat-operators-rs8f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:21 crc kubenswrapper[4787]: I1203 17:19:21.905198 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-catalog-content\") pod \"redhat-operators-rs8f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.006281 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgchs\" (UniqueName: \"kubernetes.io/projected/1b309950-19b2-4d34-b4b8-f5336e6c47f5-kube-api-access-pgchs\") pod \"redhat-operators-rs8f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.006345 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-utilities\") pod \"redhat-operators-rs8f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.006384 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-catalog-content\") pod \"redhat-operators-rs8f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.006897 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-catalog-content\") pod \"redhat-operators-rs8f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.007127 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-utilities\") pod \"redhat-operators-rs8f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.032874 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgchs\" (UniqueName: \"kubernetes.io/projected/1b309950-19b2-4d34-b4b8-f5336e6c47f5-kube-api-access-pgchs\") pod \"redhat-operators-rs8f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.116101 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.170520 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nr9fq"] Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.171809 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.180770 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nr9fq"] Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.314374 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/932d7438-a3e3-4ff4-8f7a-c70766046197-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.314431 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/932d7438-a3e3-4ff4-8f7a-c70766046197-registry-tls\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.314466 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/932d7438-a3e3-4ff4-8f7a-c70766046197-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.314482 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/932d7438-a3e3-4ff4-8f7a-c70766046197-bound-sa-token\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.314550 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.314618 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/932d7438-a3e3-4ff4-8f7a-c70766046197-registry-certificates\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.315169 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/932d7438-a3e3-4ff4-8f7a-c70766046197-trusted-ca\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.315219 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tr52m\" (UniqueName: \"kubernetes.io/projected/932d7438-a3e3-4ff4-8f7a-c70766046197-kube-api-access-tr52m\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.345166 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.416149 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/932d7438-a3e3-4ff4-8f7a-c70766046197-registry-certificates\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.416201 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/932d7438-a3e3-4ff4-8f7a-c70766046197-trusted-ca\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.416223 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tr52m\" (UniqueName: \"kubernetes.io/projected/932d7438-a3e3-4ff4-8f7a-c70766046197-kube-api-access-tr52m\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.416280 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/932d7438-a3e3-4ff4-8f7a-c70766046197-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.416311 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/932d7438-a3e3-4ff4-8f7a-c70766046197-registry-tls\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.416508 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/932d7438-a3e3-4ff4-8f7a-c70766046197-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.416528 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/932d7438-a3e3-4ff4-8f7a-c70766046197-bound-sa-token\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.417327 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/932d7438-a3e3-4ff4-8f7a-c70766046197-ca-trust-extracted\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.417945 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/932d7438-a3e3-4ff4-8f7a-c70766046197-registry-certificates\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.417976 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/932d7438-a3e3-4ff4-8f7a-c70766046197-trusted-ca\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.429689 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/932d7438-a3e3-4ff4-8f7a-c70766046197-installation-pull-secrets\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.429746 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/932d7438-a3e3-4ff4-8f7a-c70766046197-registry-tls\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.435405 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/932d7438-a3e3-4ff4-8f7a-c70766046197-bound-sa-token\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.437345 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tr52m\" (UniqueName: \"kubernetes.io/projected/932d7438-a3e3-4ff4-8f7a-c70766046197-kube-api-access-tr52m\") pod \"image-registry-66df7c8f76-nr9fq\" (UID: \"932d7438-a3e3-4ff4-8f7a-c70766046197\") " pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.513713 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.651922 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-rs8f5"] Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.789852 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kwqwl"] Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.791048 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.794389 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.807185 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kwqwl"] Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.842749 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89b68c7-da6e-447c-82f8-8662fe8efd35-catalog-content\") pod \"certified-operators-kwqwl\" (UID: \"f89b68c7-da6e-447c-82f8-8662fe8efd35\") " pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.842874 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcpgg\" (UniqueName: \"kubernetes.io/projected/f89b68c7-da6e-447c-82f8-8662fe8efd35-kube-api-access-wcpgg\") pod \"certified-operators-kwqwl\" (UID: \"f89b68c7-da6e-447c-82f8-8662fe8efd35\") " pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.842922 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89b68c7-da6e-447c-82f8-8662fe8efd35-utilities\") pod \"certified-operators-kwqwl\" (UID: \"f89b68c7-da6e-447c-82f8-8662fe8efd35\") " pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.943724 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89b68c7-da6e-447c-82f8-8662fe8efd35-utilities\") pod \"certified-operators-kwqwl\" (UID: \"f89b68c7-da6e-447c-82f8-8662fe8efd35\") " pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.943818 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89b68c7-da6e-447c-82f8-8662fe8efd35-catalog-content\") pod \"certified-operators-kwqwl\" (UID: \"f89b68c7-da6e-447c-82f8-8662fe8efd35\") " pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.943861 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcpgg\" (UniqueName: \"kubernetes.io/projected/f89b68c7-da6e-447c-82f8-8662fe8efd35-kube-api-access-wcpgg\") pod \"certified-operators-kwqwl\" (UID: \"f89b68c7-da6e-447c-82f8-8662fe8efd35\") " pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.944442 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f89b68c7-da6e-447c-82f8-8662fe8efd35-catalog-content\") pod \"certified-operators-kwqwl\" (UID: \"f89b68c7-da6e-447c-82f8-8662fe8efd35\") " pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.944442 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f89b68c7-da6e-447c-82f8-8662fe8efd35-utilities\") pod \"certified-operators-kwqwl\" (UID: \"f89b68c7-da6e-447c-82f8-8662fe8efd35\") " pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.966179 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcpgg\" (UniqueName: \"kubernetes.io/projected/f89b68c7-da6e-447c-82f8-8662fe8efd35-kube-api-access-wcpgg\") pod \"certified-operators-kwqwl\" (UID: \"f89b68c7-da6e-447c-82f8-8662fe8efd35\") " pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:22 crc kubenswrapper[4787]: I1203 17:19:22.970842 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-nr9fq"] Dec 03 17:19:22 crc kubenswrapper[4787]: W1203 17:19:22.971280 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod932d7438_a3e3_4ff4_8f7a_c70766046197.slice/crio-a72168697c61a9eecbdd23adc3707ca966e1ed660ee4ab470c368cba50c3c855 WatchSource:0}: Error finding container a72168697c61a9eecbdd23adc3707ca966e1ed660ee4ab470c368cba50c3c855: Status 404 returned error can't find the container with id a72168697c61a9eecbdd23adc3707ca966e1ed660ee4ab470c368cba50c3c855 Dec 03 17:19:23 crc kubenswrapper[4787]: I1203 17:19:23.118935 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:23 crc kubenswrapper[4787]: I1203 17:19:23.503750 4787 generic.go:334] "Generic (PLEG): container finished" podID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" containerID="f382471f7933baadc81f1f42ca12c89554df5d2f997e70a6b9ebbceaedfc2e6e" exitCode=0 Dec 03 17:19:23 crc kubenswrapper[4787]: I1203 17:19:23.504147 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs8f5" event={"ID":"1b309950-19b2-4d34-b4b8-f5336e6c47f5","Type":"ContainerDied","Data":"f382471f7933baadc81f1f42ca12c89554df5d2f997e70a6b9ebbceaedfc2e6e"} Dec 03 17:19:23 crc kubenswrapper[4787]: I1203 17:19:23.505080 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs8f5" event={"ID":"1b309950-19b2-4d34-b4b8-f5336e6c47f5","Type":"ContainerStarted","Data":"46ca95fe5526b29f4eeed28d926509d673efe896d0684e4af3d33bdd47c1bce2"} Dec 03 17:19:23 crc kubenswrapper[4787]: I1203 17:19:23.506830 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" event={"ID":"932d7438-a3e3-4ff4-8f7a-c70766046197","Type":"ContainerStarted","Data":"73d593ee88d7220f7fa62318905c6bc0ec507596d23a741689009bfc3cf9d6b3"} Dec 03 17:19:23 crc kubenswrapper[4787]: I1203 17:19:23.506873 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" event={"ID":"932d7438-a3e3-4ff4-8f7a-c70766046197","Type":"ContainerStarted","Data":"a72168697c61a9eecbdd23adc3707ca966e1ed660ee4ab470c368cba50c3c855"} Dec 03 17:19:23 crc kubenswrapper[4787]: I1203 17:19:23.507390 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:23 crc kubenswrapper[4787]: I1203 17:19:23.543862 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" podStartSLOduration=1.543841308 podStartE2EDuration="1.543841308s" podCreationTimestamp="2025-12-03 17:19:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:19:23.540945936 +0000 UTC m=+380.358416895" watchObservedRunningTime="2025-12-03 17:19:23.543841308 +0000 UTC m=+380.361312257" Dec 03 17:19:23 crc kubenswrapper[4787]: I1203 17:19:23.554775 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kwqwl"] Dec 03 17:19:23 crc kubenswrapper[4787]: W1203 17:19:23.560246 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf89b68c7_da6e_447c_82f8_8662fe8efd35.slice/crio-0440d2be3b19deebb69d502a46f644b59a01bf21f8e860f179bdd4307d56f035 WatchSource:0}: Error finding container 0440d2be3b19deebb69d502a46f644b59a01bf21f8e860f179bdd4307d56f035: Status 404 returned error can't find the container with id 0440d2be3b19deebb69d502a46f644b59a01bf21f8e860f179bdd4307d56f035 Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.191911 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-7cj8s"] Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.193968 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.195668 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.219597 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7cj8s"] Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.360921 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/280abffe-0dc9-4096-8574-74694e459f31-catalog-content\") pod \"community-operators-7cj8s\" (UID: \"280abffe-0dc9-4096-8574-74694e459f31\") " pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.360969 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/280abffe-0dc9-4096-8574-74694e459f31-utilities\") pod \"community-operators-7cj8s\" (UID: \"280abffe-0dc9-4096-8574-74694e459f31\") " pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.361093 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wztwx\" (UniqueName: \"kubernetes.io/projected/280abffe-0dc9-4096-8574-74694e459f31-kube-api-access-wztwx\") pod \"community-operators-7cj8s\" (UID: \"280abffe-0dc9-4096-8574-74694e459f31\") " pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.462544 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wztwx\" (UniqueName: \"kubernetes.io/projected/280abffe-0dc9-4096-8574-74694e459f31-kube-api-access-wztwx\") pod \"community-operators-7cj8s\" (UID: \"280abffe-0dc9-4096-8574-74694e459f31\") " pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.462622 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/280abffe-0dc9-4096-8574-74694e459f31-catalog-content\") pod \"community-operators-7cj8s\" (UID: \"280abffe-0dc9-4096-8574-74694e459f31\") " pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.462645 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/280abffe-0dc9-4096-8574-74694e459f31-utilities\") pod \"community-operators-7cj8s\" (UID: \"280abffe-0dc9-4096-8574-74694e459f31\") " pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.463087 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/280abffe-0dc9-4096-8574-74694e459f31-utilities\") pod \"community-operators-7cj8s\" (UID: \"280abffe-0dc9-4096-8574-74694e459f31\") " pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.463566 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/280abffe-0dc9-4096-8574-74694e459f31-catalog-content\") pod \"community-operators-7cj8s\" (UID: \"280abffe-0dc9-4096-8574-74694e459f31\") " pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.482530 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wztwx\" (UniqueName: \"kubernetes.io/projected/280abffe-0dc9-4096-8574-74694e459f31-kube-api-access-wztwx\") pod \"community-operators-7cj8s\" (UID: \"280abffe-0dc9-4096-8574-74694e459f31\") " pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.516601 4787 generic.go:334] "Generic (PLEG): container finished" podID="f89b68c7-da6e-447c-82f8-8662fe8efd35" containerID="c20ded06179517cbc74cfce5ccd76b00bb7cfaa50ca9ff8873ee2d7c7ac8f029" exitCode=0 Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.516680 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwqwl" event={"ID":"f89b68c7-da6e-447c-82f8-8662fe8efd35","Type":"ContainerDied","Data":"c20ded06179517cbc74cfce5ccd76b00bb7cfaa50ca9ff8873ee2d7c7ac8f029"} Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.516708 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwqwl" event={"ID":"f89b68c7-da6e-447c-82f8-8662fe8efd35","Type":"ContainerStarted","Data":"0440d2be3b19deebb69d502a46f644b59a01bf21f8e860f179bdd4307d56f035"} Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.520806 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs8f5" event={"ID":"1b309950-19b2-4d34-b4b8-f5336e6c47f5","Type":"ContainerStarted","Data":"2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776"} Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.538491 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:24 crc kubenswrapper[4787]: I1203 17:19:24.979350 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-7cj8s"] Dec 03 17:19:24 crc kubenswrapper[4787]: W1203 17:19:24.987528 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod280abffe_0dc9_4096_8574_74694e459f31.slice/crio-54f014127c2b0c38a40e7e9172e9c7485eba73aedc64510b982c0c0589584c74 WatchSource:0}: Error finding container 54f014127c2b0c38a40e7e9172e9c7485eba73aedc64510b982c0c0589584c74: Status 404 returned error can't find the container with id 54f014127c2b0c38a40e7e9172e9c7485eba73aedc64510b982c0c0589584c74 Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.198382 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-5mwph"] Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.200998 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.203475 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.210149 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mwph"] Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.378052 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rvmpf\" (UniqueName: \"kubernetes.io/projected/4ffe2b99-3c93-493c-899f-be8d2e6a65b7-kube-api-access-rvmpf\") pod \"redhat-marketplace-5mwph\" (UID: \"4ffe2b99-3c93-493c-899f-be8d2e6a65b7\") " pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.378500 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ffe2b99-3c93-493c-899f-be8d2e6a65b7-catalog-content\") pod \"redhat-marketplace-5mwph\" (UID: \"4ffe2b99-3c93-493c-899f-be8d2e6a65b7\") " pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.378562 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ffe2b99-3c93-493c-899f-be8d2e6a65b7-utilities\") pod \"redhat-marketplace-5mwph\" (UID: \"4ffe2b99-3c93-493c-899f-be8d2e6a65b7\") " pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.479447 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ffe2b99-3c93-493c-899f-be8d2e6a65b7-catalog-content\") pod \"redhat-marketplace-5mwph\" (UID: \"4ffe2b99-3c93-493c-899f-be8d2e6a65b7\") " pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.480054 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ffe2b99-3c93-493c-899f-be8d2e6a65b7-utilities\") pod \"redhat-marketplace-5mwph\" (UID: \"4ffe2b99-3c93-493c-899f-be8d2e6a65b7\") " pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.480072 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4ffe2b99-3c93-493c-899f-be8d2e6a65b7-catalog-content\") pod \"redhat-marketplace-5mwph\" (UID: \"4ffe2b99-3c93-493c-899f-be8d2e6a65b7\") " pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.480347 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4ffe2b99-3c93-493c-899f-be8d2e6a65b7-utilities\") pod \"redhat-marketplace-5mwph\" (UID: \"4ffe2b99-3c93-493c-899f-be8d2e6a65b7\") " pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.480512 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rvmpf\" (UniqueName: \"kubernetes.io/projected/4ffe2b99-3c93-493c-899f-be8d2e6a65b7-kube-api-access-rvmpf\") pod \"redhat-marketplace-5mwph\" (UID: \"4ffe2b99-3c93-493c-899f-be8d2e6a65b7\") " pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.499613 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rvmpf\" (UniqueName: \"kubernetes.io/projected/4ffe2b99-3c93-493c-899f-be8d2e6a65b7-kube-api-access-rvmpf\") pod \"redhat-marketplace-5mwph\" (UID: \"4ffe2b99-3c93-493c-899f-be8d2e6a65b7\") " pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.529335 4787 generic.go:334] "Generic (PLEG): container finished" podID="280abffe-0dc9-4096-8574-74694e459f31" containerID="893148b47384d5df9b4c3a2858e802263c867d637b36481ba0ed7efd7b86995a" exitCode=0 Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.529386 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cj8s" event={"ID":"280abffe-0dc9-4096-8574-74694e459f31","Type":"ContainerDied","Data":"893148b47384d5df9b4c3a2858e802263c867d637b36481ba0ed7efd7b86995a"} Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.529433 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cj8s" event={"ID":"280abffe-0dc9-4096-8574-74694e459f31","Type":"ContainerStarted","Data":"54f014127c2b0c38a40e7e9172e9c7485eba73aedc64510b982c0c0589584c74"} Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.531724 4787 generic.go:334] "Generic (PLEG): container finished" podID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" containerID="2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776" exitCode=0 Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.531839 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs8f5" event={"ID":"1b309950-19b2-4d34-b4b8-f5336e6c47f5","Type":"ContainerDied","Data":"2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776"} Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.534982 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwqwl" event={"ID":"f89b68c7-da6e-447c-82f8-8662fe8efd35","Type":"ContainerStarted","Data":"e85c160c27fe2d892c151d709063f88e67b14c17e324bb2e3bb4f0d3784d5c3e"} Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.540646 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:25 crc kubenswrapper[4787]: I1203 17:19:25.938100 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-5mwph"] Dec 03 17:19:26 crc kubenswrapper[4787]: I1203 17:19:26.540723 4787 generic.go:334] "Generic (PLEG): container finished" podID="280abffe-0dc9-4096-8574-74694e459f31" containerID="cc3b47d44f9412bd5306125e5ff1749544256cd1daa306079a5978359bda1148" exitCode=0 Dec 03 17:19:26 crc kubenswrapper[4787]: I1203 17:19:26.540916 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cj8s" event={"ID":"280abffe-0dc9-4096-8574-74694e459f31","Type":"ContainerDied","Data":"cc3b47d44f9412bd5306125e5ff1749544256cd1daa306079a5978359bda1148"} Dec 03 17:19:26 crc kubenswrapper[4787]: I1203 17:19:26.544758 4787 generic.go:334] "Generic (PLEG): container finished" podID="4ffe2b99-3c93-493c-899f-be8d2e6a65b7" containerID="3eed7be860f64aeca595e074b88fcfbbade4cf916f2eb03339e3edb43fa3ea00" exitCode=0 Dec 03 17:19:26 crc kubenswrapper[4787]: I1203 17:19:26.544844 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mwph" event={"ID":"4ffe2b99-3c93-493c-899f-be8d2e6a65b7","Type":"ContainerDied","Data":"3eed7be860f64aeca595e074b88fcfbbade4cf916f2eb03339e3edb43fa3ea00"} Dec 03 17:19:26 crc kubenswrapper[4787]: I1203 17:19:26.544872 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mwph" event={"ID":"4ffe2b99-3c93-493c-899f-be8d2e6a65b7","Type":"ContainerStarted","Data":"7934d9d4173c691c8b462019e3645eac732b08b377ee116057eaade35586a8c5"} Dec 03 17:19:26 crc kubenswrapper[4787]: I1203 17:19:26.547875 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs8f5" event={"ID":"1b309950-19b2-4d34-b4b8-f5336e6c47f5","Type":"ContainerStarted","Data":"d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04"} Dec 03 17:19:26 crc kubenswrapper[4787]: I1203 17:19:26.552223 4787 generic.go:334] "Generic (PLEG): container finished" podID="f89b68c7-da6e-447c-82f8-8662fe8efd35" containerID="e85c160c27fe2d892c151d709063f88e67b14c17e324bb2e3bb4f0d3784d5c3e" exitCode=0 Dec 03 17:19:26 crc kubenswrapper[4787]: I1203 17:19:26.552263 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwqwl" event={"ID":"f89b68c7-da6e-447c-82f8-8662fe8efd35","Type":"ContainerDied","Data":"e85c160c27fe2d892c151d709063f88e67b14c17e324bb2e3bb4f0d3784d5c3e"} Dec 03 17:19:26 crc kubenswrapper[4787]: I1203 17:19:26.608553 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-rs8f5" podStartSLOduration=3.172328776 podStartE2EDuration="5.608534923s" podCreationTimestamp="2025-12-03 17:19:21 +0000 UTC" firstStartedPulling="2025-12-03 17:19:23.506130353 +0000 UTC m=+380.323601312" lastFinishedPulling="2025-12-03 17:19:25.9423365 +0000 UTC m=+382.759807459" observedRunningTime="2025-12-03 17:19:26.605753295 +0000 UTC m=+383.423224254" watchObservedRunningTime="2025-12-03 17:19:26.608534923 +0000 UTC m=+383.426005882" Dec 03 17:19:28 crc kubenswrapper[4787]: I1203 17:19:28.571285 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kwqwl" event={"ID":"f89b68c7-da6e-447c-82f8-8662fe8efd35","Type":"ContainerStarted","Data":"16d336d8aada7fa09ad102adfb3d686a17d9bd599c074a7f38f9106fdf23ad33"} Dec 03 17:19:28 crc kubenswrapper[4787]: I1203 17:19:28.578934 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-7cj8s" event={"ID":"280abffe-0dc9-4096-8574-74694e459f31","Type":"ContainerStarted","Data":"0bcbdb15eefe1693166944d4183d6d48db39c0dd537f6b8f6d618334a31c848b"} Dec 03 17:19:28 crc kubenswrapper[4787]: I1203 17:19:28.582299 4787 generic.go:334] "Generic (PLEG): container finished" podID="4ffe2b99-3c93-493c-899f-be8d2e6a65b7" containerID="cf6a8fe7e1edc1e4b95d58a1bd5434686fe33b600e82fa6ee24f18debdf84edb" exitCode=0 Dec 03 17:19:28 crc kubenswrapper[4787]: I1203 17:19:28.582347 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mwph" event={"ID":"4ffe2b99-3c93-493c-899f-be8d2e6a65b7","Type":"ContainerDied","Data":"cf6a8fe7e1edc1e4b95d58a1bd5434686fe33b600e82fa6ee24f18debdf84edb"} Dec 03 17:19:28 crc kubenswrapper[4787]: I1203 17:19:28.589800 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kwqwl" podStartSLOduration=4.114596134 podStartE2EDuration="6.589784672s" podCreationTimestamp="2025-12-03 17:19:22 +0000 UTC" firstStartedPulling="2025-12-03 17:19:24.518868172 +0000 UTC m=+381.336339131" lastFinishedPulling="2025-12-03 17:19:26.99405671 +0000 UTC m=+383.811527669" observedRunningTime="2025-12-03 17:19:28.587609341 +0000 UTC m=+385.405080300" watchObservedRunningTime="2025-12-03 17:19:28.589784672 +0000 UTC m=+385.407255631" Dec 03 17:19:28 crc kubenswrapper[4787]: I1203 17:19:28.629185 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-7cj8s" podStartSLOduration=3.2426076679999998 podStartE2EDuration="4.629168154s" podCreationTimestamp="2025-12-03 17:19:24 +0000 UTC" firstStartedPulling="2025-12-03 17:19:25.531243011 +0000 UTC m=+382.348713970" lastFinishedPulling="2025-12-03 17:19:26.917803497 +0000 UTC m=+383.735274456" observedRunningTime="2025-12-03 17:19:28.62545816 +0000 UTC m=+385.442929119" watchObservedRunningTime="2025-12-03 17:19:28.629168154 +0000 UTC m=+385.446639113" Dec 03 17:19:31 crc kubenswrapper[4787]: I1203 17:19:31.600265 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-5mwph" event={"ID":"4ffe2b99-3c93-493c-899f-be8d2e6a65b7","Type":"ContainerStarted","Data":"18825ffa06d584af029fb93282b0aed505d02f828be5a75e4beae9ab58a6c370"} Dec 03 17:19:32 crc kubenswrapper[4787]: I1203 17:19:32.116401 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:32 crc kubenswrapper[4787]: I1203 17:19:32.116862 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:32 crc kubenswrapper[4787]: I1203 17:19:32.166447 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:32 crc kubenswrapper[4787]: I1203 17:19:32.184108 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-5mwph" podStartSLOduration=3.186808442 podStartE2EDuration="7.184089974s" podCreationTimestamp="2025-12-03 17:19:25 +0000 UTC" firstStartedPulling="2025-12-03 17:19:26.546177022 +0000 UTC m=+383.363647981" lastFinishedPulling="2025-12-03 17:19:30.543458554 +0000 UTC m=+387.360929513" observedRunningTime="2025-12-03 17:19:31.625226022 +0000 UTC m=+388.442696981" watchObservedRunningTime="2025-12-03 17:19:32.184089974 +0000 UTC m=+389.001560933" Dec 03 17:19:32 crc kubenswrapper[4787]: I1203 17:19:32.653145 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 17:19:33 crc kubenswrapper[4787]: I1203 17:19:33.119591 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:33 crc kubenswrapper[4787]: I1203 17:19:33.119866 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:33 crc kubenswrapper[4787]: I1203 17:19:33.155699 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:33 crc kubenswrapper[4787]: I1203 17:19:33.648306 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kwqwl" Dec 03 17:19:34 crc kubenswrapper[4787]: I1203 17:19:34.539701 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:34 crc kubenswrapper[4787]: I1203 17:19:34.539770 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:34 crc kubenswrapper[4787]: I1203 17:19:34.578484 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:34 crc kubenswrapper[4787]: I1203 17:19:34.657481 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-7cj8s" Dec 03 17:19:35 crc kubenswrapper[4787]: I1203 17:19:35.541395 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:35 crc kubenswrapper[4787]: I1203 17:19:35.541493 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:35 crc kubenswrapper[4787]: I1203 17:19:35.580187 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:35 crc kubenswrapper[4787]: I1203 17:19:35.658746 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-5mwph" Dec 03 17:19:42 crc kubenswrapper[4787]: I1203 17:19:42.522265 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-nr9fq" Dec 03 17:19:42 crc kubenswrapper[4787]: I1203 17:19:42.598708 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dw4qj"] Dec 03 17:19:48 crc kubenswrapper[4787]: I1203 17:19:48.989354 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:19:48 crc kubenswrapper[4787]: I1203 17:19:48.990985 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:20:07 crc kubenswrapper[4787]: I1203 17:20:07.667250 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" podUID="2df4d7cb-f9bb-4ece-9b95-36027a01f017" containerName="registry" containerID="cri-o://7a682029f18a58605dc8172e19f6760b5ac934129a032b613329e7e395623eaa" gracePeriod=30 Dec 03 17:20:07 crc kubenswrapper[4787]: I1203 17:20:07.813573 4787 generic.go:334] "Generic (PLEG): container finished" podID="2df4d7cb-f9bb-4ece-9b95-36027a01f017" containerID="7a682029f18a58605dc8172e19f6760b5ac934129a032b613329e7e395623eaa" exitCode=0 Dec 03 17:20:07 crc kubenswrapper[4787]: I1203 17:20:07.813620 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" event={"ID":"2df4d7cb-f9bb-4ece-9b95-36027a01f017","Type":"ContainerDied","Data":"7a682029f18a58605dc8172e19f6760b5ac934129a032b613329e7e395623eaa"} Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.122493 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.164237 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2df4d7cb-f9bb-4ece-9b95-36027a01f017-installation-pull-secrets\") pod \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.164288 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-certificates\") pod \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.164333 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-tls\") pod \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.164360 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-trusted-ca\") pod \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.164375 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jvwvq\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-kube-api-access-jvwvq\") pod \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.164390 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2df4d7cb-f9bb-4ece-9b95-36027a01f017-ca-trust-extracted\") pod \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.164406 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-bound-sa-token\") pod \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.164529 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\" (UID: \"2df4d7cb-f9bb-4ece-9b95-36027a01f017\") " Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.165848 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "2df4d7cb-f9bb-4ece-9b95-36027a01f017" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.166658 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "2df4d7cb-f9bb-4ece-9b95-36027a01f017" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.171350 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-kube-api-access-jvwvq" (OuterVolumeSpecName: "kube-api-access-jvwvq") pod "2df4d7cb-f9bb-4ece-9b95-36027a01f017" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017"). InnerVolumeSpecName "kube-api-access-jvwvq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.171899 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "2df4d7cb-f9bb-4ece-9b95-36027a01f017" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.172614 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "2df4d7cb-f9bb-4ece-9b95-36027a01f017" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.173657 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2df4d7cb-f9bb-4ece-9b95-36027a01f017-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "2df4d7cb-f9bb-4ece-9b95-36027a01f017" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.180706 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "2df4d7cb-f9bb-4ece-9b95-36027a01f017" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.181439 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2df4d7cb-f9bb-4ece-9b95-36027a01f017-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "2df4d7cb-f9bb-4ece-9b95-36027a01f017" (UID: "2df4d7cb-f9bb-4ece-9b95-36027a01f017"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.265394 4787 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/2df4d7cb-f9bb-4ece-9b95-36027a01f017-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.265434 4787 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.265445 4787 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.265457 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2df4d7cb-f9bb-4ece-9b95-36027a01f017-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.265467 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jvwvq\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-kube-api-access-jvwvq\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.265476 4787 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/2df4d7cb-f9bb-4ece-9b95-36027a01f017-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.265484 4787 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2df4d7cb-f9bb-4ece-9b95-36027a01f017-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.822889 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" event={"ID":"2df4d7cb-f9bb-4ece-9b95-36027a01f017","Type":"ContainerDied","Data":"70b8c57ca75541845296c13338ab4de34f045220318a2938e6e997e16eb6883d"} Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.822966 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-dw4qj" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.822983 4787 scope.go:117] "RemoveContainer" containerID="7a682029f18a58605dc8172e19f6760b5ac934129a032b613329e7e395623eaa" Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.881012 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dw4qj"] Dec 03 17:20:08 crc kubenswrapper[4787]: I1203 17:20:08.891915 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-dw4qj"] Dec 03 17:20:09 crc kubenswrapper[4787]: I1203 17:20:09.775756 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2df4d7cb-f9bb-4ece-9b95-36027a01f017" path="/var/lib/kubelet/pods/2df4d7cb-f9bb-4ece-9b95-36027a01f017/volumes" Dec 03 17:20:18 crc kubenswrapper[4787]: I1203 17:20:18.989416 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:20:18 crc kubenswrapper[4787]: I1203 17:20:18.990184 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:20:18 crc kubenswrapper[4787]: I1203 17:20:18.990236 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:20:18 crc kubenswrapper[4787]: I1203 17:20:18.990929 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d126a06f7bcbd1acb6bb4b36067339af8c76b0636d70dda2e29f3ff99ca18e1c"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:20:18 crc kubenswrapper[4787]: I1203 17:20:18.990995 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://d126a06f7bcbd1acb6bb4b36067339af8c76b0636d70dda2e29f3ff99ca18e1c" gracePeriod=600 Dec 03 17:20:19 crc kubenswrapper[4787]: I1203 17:20:19.889805 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="d126a06f7bcbd1acb6bb4b36067339af8c76b0636d70dda2e29f3ff99ca18e1c" exitCode=0 Dec 03 17:20:19 crc kubenswrapper[4787]: I1203 17:20:19.889928 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"d126a06f7bcbd1acb6bb4b36067339af8c76b0636d70dda2e29f3ff99ca18e1c"} Dec 03 17:20:19 crc kubenswrapper[4787]: I1203 17:20:19.890582 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"971d7c10a947bd7777cac22b6df04c66ed8ba415706fe05793b9542a5980006b"} Dec 03 17:20:19 crc kubenswrapper[4787]: I1203 17:20:19.890642 4787 scope.go:117] "RemoveContainer" containerID="72ee03a0c7209854fe589caf9291f8556e4f133614cc5564e57109e41ef09f26" Dec 03 17:22:48 crc kubenswrapper[4787]: I1203 17:22:48.989948 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:22:48 crc kubenswrapper[4787]: I1203 17:22:48.990583 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:23:18 crc kubenswrapper[4787]: I1203 17:23:18.990602 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:23:18 crc kubenswrapper[4787]: I1203 17:23:18.991619 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:23:48 crc kubenswrapper[4787]: I1203 17:23:48.989866 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:23:48 crc kubenswrapper[4787]: I1203 17:23:48.990469 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:23:48 crc kubenswrapper[4787]: I1203 17:23:48.990537 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:23:48 crc kubenswrapper[4787]: I1203 17:23:48.992241 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"971d7c10a947bd7777cac22b6df04c66ed8ba415706fe05793b9542a5980006b"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:23:48 crc kubenswrapper[4787]: I1203 17:23:48.992420 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://971d7c10a947bd7777cac22b6df04c66ed8ba415706fe05793b9542a5980006b" gracePeriod=600 Dec 03 17:23:50 crc kubenswrapper[4787]: I1203 17:23:50.129082 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="971d7c10a947bd7777cac22b6df04c66ed8ba415706fe05793b9542a5980006b" exitCode=0 Dec 03 17:23:50 crc kubenswrapper[4787]: I1203 17:23:50.129268 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"971d7c10a947bd7777cac22b6df04c66ed8ba415706fe05793b9542a5980006b"} Dec 03 17:23:50 crc kubenswrapper[4787]: I1203 17:23:50.129596 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"49596b58389f8b6bbf73d259649cd03b2e93ae7e21af7479d7bc4161137e279b"} Dec 03 17:23:50 crc kubenswrapper[4787]: I1203 17:23:50.129620 4787 scope.go:117] "RemoveContainer" containerID="d126a06f7bcbd1acb6bb4b36067339af8c76b0636d70dda2e29f3ff99ca18e1c" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.214148 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc"] Dec 03 17:24:58 crc kubenswrapper[4787]: E1203 17:24:58.215389 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2df4d7cb-f9bb-4ece-9b95-36027a01f017" containerName="registry" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.215413 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2df4d7cb-f9bb-4ece-9b95-36027a01f017" containerName="registry" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.215593 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2df4d7cb-f9bb-4ece-9b95-36027a01f017" containerName="registry" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.216762 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.220000 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.285567 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc"] Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.309495 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwz9b\" (UniqueName: \"kubernetes.io/projected/e3390472-0d01-4c86-8756-cca8bc29a289-kube-api-access-dwz9b\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.309593 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.309659 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.410964 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.411062 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwz9b\" (UniqueName: \"kubernetes.io/projected/e3390472-0d01-4c86-8756-cca8bc29a289-kube-api-access-dwz9b\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.411096 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.411656 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-bundle\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.411992 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-util\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.440885 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwz9b\" (UniqueName: \"kubernetes.io/projected/e3390472-0d01-4c86-8756-cca8bc29a289-kube-api-access-dwz9b\") pod \"6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.539629 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:24:58 crc kubenswrapper[4787]: I1203 17:24:58.749661 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc"] Dec 03 17:24:59 crc kubenswrapper[4787]: I1203 17:24:59.534333 4787 generic.go:334] "Generic (PLEG): container finished" podID="e3390472-0d01-4c86-8756-cca8bc29a289" containerID="d6789a9eb5404213621b5246d6046e8edadb3351e79aa4f7f5accdcda748b26e" exitCode=0 Dec 03 17:24:59 crc kubenswrapper[4787]: I1203 17:24:59.534391 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" event={"ID":"e3390472-0d01-4c86-8756-cca8bc29a289","Type":"ContainerDied","Data":"d6789a9eb5404213621b5246d6046e8edadb3351e79aa4f7f5accdcda748b26e"} Dec 03 17:24:59 crc kubenswrapper[4787]: I1203 17:24:59.534642 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" event={"ID":"e3390472-0d01-4c86-8756-cca8bc29a289","Type":"ContainerStarted","Data":"87dc00042321cc1a2d8e756243650c05efa00614c341e7dc4979021e9c348048"} Dec 03 17:24:59 crc kubenswrapper[4787]: I1203 17:24:59.536888 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:25:01 crc kubenswrapper[4787]: I1203 17:25:01.546934 4787 generic.go:334] "Generic (PLEG): container finished" podID="e3390472-0d01-4c86-8756-cca8bc29a289" containerID="fd2e757088c2829c3c361fab166148e850d368f9da328dcb44402d831f95713b" exitCode=0 Dec 03 17:25:01 crc kubenswrapper[4787]: I1203 17:25:01.546988 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" event={"ID":"e3390472-0d01-4c86-8756-cca8bc29a289","Type":"ContainerDied","Data":"fd2e757088c2829c3c361fab166148e850d368f9da328dcb44402d831f95713b"} Dec 03 17:25:02 crc kubenswrapper[4787]: I1203 17:25:02.557094 4787 generic.go:334] "Generic (PLEG): container finished" podID="e3390472-0d01-4c86-8756-cca8bc29a289" containerID="480b2499f036d551beb6538a850e15f5d16daee14223d8bcb371dabbb05c50e2" exitCode=0 Dec 03 17:25:02 crc kubenswrapper[4787]: I1203 17:25:02.557548 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" event={"ID":"e3390472-0d01-4c86-8756-cca8bc29a289","Type":"ContainerDied","Data":"480b2499f036d551beb6538a850e15f5d16daee14223d8bcb371dabbb05c50e2"} Dec 03 17:25:03 crc kubenswrapper[4787]: I1203 17:25:03.862678 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:25:03 crc kubenswrapper[4787]: I1203 17:25:03.984682 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-util\") pod \"e3390472-0d01-4c86-8756-cca8bc29a289\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " Dec 03 17:25:03 crc kubenswrapper[4787]: I1203 17:25:03.984824 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-bundle\") pod \"e3390472-0d01-4c86-8756-cca8bc29a289\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " Dec 03 17:25:03 crc kubenswrapper[4787]: I1203 17:25:03.984928 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwz9b\" (UniqueName: \"kubernetes.io/projected/e3390472-0d01-4c86-8756-cca8bc29a289-kube-api-access-dwz9b\") pod \"e3390472-0d01-4c86-8756-cca8bc29a289\" (UID: \"e3390472-0d01-4c86-8756-cca8bc29a289\") " Dec 03 17:25:03 crc kubenswrapper[4787]: I1203 17:25:03.989588 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-bundle" (OuterVolumeSpecName: "bundle") pod "e3390472-0d01-4c86-8756-cca8bc29a289" (UID: "e3390472-0d01-4c86-8756-cca8bc29a289"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:25:03 crc kubenswrapper[4787]: I1203 17:25:03.991889 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3390472-0d01-4c86-8756-cca8bc29a289-kube-api-access-dwz9b" (OuterVolumeSpecName: "kube-api-access-dwz9b") pod "e3390472-0d01-4c86-8756-cca8bc29a289" (UID: "e3390472-0d01-4c86-8756-cca8bc29a289"). InnerVolumeSpecName "kube-api-access-dwz9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:25:04 crc kubenswrapper[4787]: I1203 17:25:04.015055 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-util" (OuterVolumeSpecName: "util") pod "e3390472-0d01-4c86-8756-cca8bc29a289" (UID: "e3390472-0d01-4c86-8756-cca8bc29a289"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:25:04 crc kubenswrapper[4787]: I1203 17:25:04.085961 4787 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:04 crc kubenswrapper[4787]: I1203 17:25:04.085999 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwz9b\" (UniqueName: \"kubernetes.io/projected/e3390472-0d01-4c86-8756-cca8bc29a289-kube-api-access-dwz9b\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:04 crc kubenswrapper[4787]: I1203 17:25:04.086012 4787 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/e3390472-0d01-4c86-8756-cca8bc29a289-util\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:04 crc kubenswrapper[4787]: I1203 17:25:04.572994 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" event={"ID":"e3390472-0d01-4c86-8756-cca8bc29a289","Type":"ContainerDied","Data":"87dc00042321cc1a2d8e756243650c05efa00614c341e7dc4979021e9c348048"} Dec 03 17:25:04 crc kubenswrapper[4787]: I1203 17:25:04.573064 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87dc00042321cc1a2d8e756243650c05efa00614c341e7dc4979021e9c348048" Dec 03 17:25:04 crc kubenswrapper[4787]: I1203 17:25:04.573118 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.262206 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bx8nf"] Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.262946 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovn-controller" containerID="cri-o://5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e" gracePeriod=30 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.262986 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="nbdb" containerID="cri-o://cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f" gracePeriod=30 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.263056 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516" gracePeriod=30 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.262970 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="northd" containerID="cri-o://24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df" gracePeriod=30 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.263072 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="sbdb" containerID="cri-o://0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c" gracePeriod=30 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.263091 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="kube-rbac-proxy-node" containerID="cri-o://b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3" gracePeriod=30 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.263148 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovn-acl-logging" containerID="cri-o://ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4" gracePeriod=30 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.306676 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" containerID="cri-o://08b8f23afaf8a16b6613bf1e8f2bfea737d61a693f5a6f6c3c2213cf3202a093" gracePeriod=30 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.604109 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xcz4f_bd5617c0-aa4f-4f21-a131-15af831725ec/kube-multus/2.log" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.604845 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xcz4f_bd5617c0-aa4f-4f21-a131-15af831725ec/kube-multus/1.log" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.604925 4787 generic.go:334] "Generic (PLEG): container finished" podID="bd5617c0-aa4f-4f21-a131-15af831725ec" containerID="d24796ebc7e29b1d6ddb686edefd44466c78b350cb28a0f36ea374203323649a" exitCode=2 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.605047 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xcz4f" event={"ID":"bd5617c0-aa4f-4f21-a131-15af831725ec","Type":"ContainerDied","Data":"d24796ebc7e29b1d6ddb686edefd44466c78b350cb28a0f36ea374203323649a"} Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.605106 4787 scope.go:117] "RemoveContainer" containerID="59da80fc47dd28be64dc70de088c9eaf83ed5a8136fd490b12d08aca6c6db4f8" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.607441 4787 scope.go:117] "RemoveContainer" containerID="d24796ebc7e29b1d6ddb686edefd44466c78b350cb28a0f36ea374203323649a" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.607718 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-xcz4f_openshift-multus(bd5617c0-aa4f-4f21-a131-15af831725ec)\"" pod="openshift-multus/multus-xcz4f" podUID="bd5617c0-aa4f-4f21-a131-15af831725ec" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.609336 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/3.log" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.612497 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovn-acl-logging/0.log" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.613067 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovn-controller/0.log" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614341 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="08b8f23afaf8a16b6613bf1e8f2bfea737d61a693f5a6f6c3c2213cf3202a093" exitCode=0 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614374 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c" exitCode=0 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614385 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f" exitCode=0 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614395 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df" exitCode=0 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614405 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516" exitCode=0 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614413 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3" exitCode=0 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614424 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4" exitCode=143 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614433 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerID="5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e" exitCode=143 Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614432 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"08b8f23afaf8a16b6613bf1e8f2bfea737d61a693f5a6f6c3c2213cf3202a093"} Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614479 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c"} Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614498 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f"} Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614518 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df"} Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614534 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516"} Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614549 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3"} Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614564 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4"} Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.614579 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e"} Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.644173 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovnkube-controller/3.log" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.645444 4787 scope.go:117] "RemoveContainer" containerID="d9e1f286dffc0f40fb247e6cbf4ff9d2353f2995319969bbc6b377451b6482d6" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.645983 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovn-acl-logging/0.log" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.646596 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovn-controller/0.log" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.647065 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702323 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-swdrj"] Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702619 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovn-acl-logging" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702636 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovn-acl-logging" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702650 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702672 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702679 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="kubecfg-setup" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702685 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="kubecfg-setup" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702694 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="sbdb" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702699 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="sbdb" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702708 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3390472-0d01-4c86-8756-cca8bc29a289" containerName="extract" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702714 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3390472-0d01-4c86-8756-cca8bc29a289" containerName="extract" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702721 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702727 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702753 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702763 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702773 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="kube-rbac-proxy-node" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702780 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="kube-rbac-proxy-node" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702789 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="northd" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702796 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="northd" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702808 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="nbdb" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702834 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="nbdb" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702844 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702850 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702860 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3390472-0d01-4c86-8756-cca8bc29a289" containerName="pull" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702867 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3390472-0d01-4c86-8756-cca8bc29a289" containerName="pull" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702877 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3390472-0d01-4c86-8756-cca8bc29a289" containerName="util" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702884 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3390472-0d01-4c86-8756-cca8bc29a289" containerName="util" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.702914 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovn-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.702923 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovn-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703088 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703097 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3390472-0d01-4c86-8756-cca8bc29a289" containerName="extract" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703105 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="northd" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703111 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703118 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="nbdb" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703598 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703643 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="kube-rbac-proxy-ovn-metrics" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703651 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703659 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovn-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703666 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="sbdb" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703672 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovn-acl-logging" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703678 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="kube-rbac-proxy-node" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.703889 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.703899 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.704048 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: E1203 17:25:09.704178 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.704218 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" containerName="ovnkube-controller" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.706152 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.772983 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-config\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.773381 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-node-log\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.773475 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-slash\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.773618 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-systemd\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.773747 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-script-lib\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.773475 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-node-log" (OuterVolumeSpecName: "node-log") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.773487 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.773494 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-slash" (OuterVolumeSpecName: "host-slash") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.774026 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.774131 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-ovn\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.774232 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-log-socket\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.774347 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovn-node-metrics-cert\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.774864 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-var-lib-openvswitch\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.774267 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.774274 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-log-socket" (OuterVolumeSpecName: "log-socket") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.774929 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.774956 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-kubelet\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775075 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-netd\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775102 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-ovn-kubernetes\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775134 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-netns\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775158 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-bin\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775197 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-systemd-units\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775226 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5k4bd\" (UniqueName: \"kubernetes.io/projected/c3f5a989-6421-4bc7-a516-b37c825d8bbc-kube-api-access-5k4bd\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775253 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-openvswitch\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775285 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-etc-openvswitch\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775312 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-env-overrides\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775354 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-var-lib-cni-networks-ovn-kubernetes\") pod \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\" (UID: \"c3f5a989-6421-4bc7-a516-b37c825d8bbc\") " Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775362 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775383 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775395 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775397 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775372 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775421 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775445 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775447 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775487 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.775868 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.776213 4787 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.776300 4787 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.776377 4787 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.776447 4787 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.776523 4787 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.776648 4787 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.776728 4787 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-node-log\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.776799 4787 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-slash\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.776869 4787 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.776937 4787 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.777004 4787 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-log-socket\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.777104 4787 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.777170 4787 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.777238 4787 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.777304 4787 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.777371 4787 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.777449 4787 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.779793 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.779877 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3f5a989-6421-4bc7-a516-b37c825d8bbc-kube-api-access-5k4bd" (OuterVolumeSpecName: "kube-api-access-5k4bd") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "kube-api-access-5k4bd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.788591 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "c3f5a989-6421-4bc7-a516-b37c825d8bbc" (UID: "c3f5a989-6421-4bc7-a516-b37c825d8bbc"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.878734 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68802a81-1754-4f82-a98f-5f45fc7276d8-ovn-node-metrics-cert\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.879002 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-run-openvswitch\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.879143 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.879275 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-kubelet\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.879377 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-run-ovn\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.879474 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/68802a81-1754-4f82-a98f-5f45fc7276d8-ovnkube-script-lib\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.879573 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-cni-netd\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.879676 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-etc-openvswitch\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.879774 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-run-ovn-kubernetes\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.879865 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-run-netns\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.879953 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68802a81-1754-4f82-a98f-5f45fc7276d8-ovnkube-config\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880061 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-var-lib-openvswitch\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880172 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-slash\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880269 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-run-systemd\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880356 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-systemd-units\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880446 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-node-log\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880533 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-cni-bin\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880629 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-log-socket\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880715 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8vdbk\" (UniqueName: \"kubernetes.io/projected/68802a81-1754-4f82-a98f-5f45fc7276d8-kube-api-access-8vdbk\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880828 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68802a81-1754-4f82-a98f-5f45fc7276d8-env-overrides\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880897 4787 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/c3f5a989-6421-4bc7-a516-b37c825d8bbc-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880916 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c3f5a989-6421-4bc7-a516-b37c825d8bbc-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.880934 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5k4bd\" (UniqueName: \"kubernetes.io/projected/c3f5a989-6421-4bc7-a516-b37c825d8bbc-kube-api-access-5k4bd\") on node \"crc\" DevicePath \"\"" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982534 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-node-log\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982601 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-cni-bin\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982651 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-log-socket\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982683 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8vdbk\" (UniqueName: \"kubernetes.io/projected/68802a81-1754-4f82-a98f-5f45fc7276d8-kube-api-access-8vdbk\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982718 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68802a81-1754-4f82-a98f-5f45fc7276d8-env-overrides\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982753 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68802a81-1754-4f82-a98f-5f45fc7276d8-ovn-node-metrics-cert\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982791 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-run-openvswitch\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982826 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982859 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-kubelet\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982897 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-run-ovn\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982934 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/68802a81-1754-4f82-a98f-5f45fc7276d8-ovnkube-script-lib\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.982969 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-cni-netd\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983007 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-etc-openvswitch\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983064 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-run-ovn-kubernetes\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983095 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-run-netns\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983130 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68802a81-1754-4f82-a98f-5f45fc7276d8-ovnkube-config\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983162 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-var-lib-openvswitch\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983200 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-slash\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983236 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-run-systemd\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983269 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-systemd-units\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983372 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-systemd-units\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983431 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-node-log\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983475 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-cni-bin\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.983515 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-log-socket\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.984813 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/68802a81-1754-4f82-a98f-5f45fc7276d8-env-overrides\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.984886 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-cni-netd\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.985500 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-var-lib-openvswitch\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.985557 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-slash\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.985597 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-run-systemd\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.985637 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-run-ovn-kubernetes\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.985685 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-etc-openvswitch\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.985760 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-kubelet\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.985807 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-run-openvswitch\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.985847 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.985888 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-run-ovn\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.985930 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/68802a81-1754-4f82-a98f-5f45fc7276d8-host-run-netns\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.986208 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/68802a81-1754-4f82-a98f-5f45fc7276d8-ovnkube-config\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.986882 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/68802a81-1754-4f82-a98f-5f45fc7276d8-ovnkube-script-lib\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:09 crc kubenswrapper[4787]: I1203 17:25:09.988485 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/68802a81-1754-4f82-a98f-5f45fc7276d8-ovn-node-metrics-cert\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.020716 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8vdbk\" (UniqueName: \"kubernetes.io/projected/68802a81-1754-4f82-a98f-5f45fc7276d8-kube-api-access-8vdbk\") pod \"ovnkube-node-swdrj\" (UID: \"68802a81-1754-4f82-a98f-5f45fc7276d8\") " pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.021541 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.620997 4787 generic.go:334] "Generic (PLEG): container finished" podID="68802a81-1754-4f82-a98f-5f45fc7276d8" containerID="492d9a717d70fa2f5d22c8e3d25bf5a5db1b001ad1467ee71b79c7c525d13348" exitCode=0 Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.621066 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" event={"ID":"68802a81-1754-4f82-a98f-5f45fc7276d8","Type":"ContainerDied","Data":"492d9a717d70fa2f5d22c8e3d25bf5a5db1b001ad1467ee71b79c7c525d13348"} Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.621354 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" event={"ID":"68802a81-1754-4f82-a98f-5f45fc7276d8","Type":"ContainerStarted","Data":"e56d4a028c395690334c228f4106c019b607632f00419ad093507702f85ed579"} Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.625332 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xcz4f_bd5617c0-aa4f-4f21-a131-15af831725ec/kube-multus/2.log" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.630090 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovn-acl-logging/0.log" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.630590 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-bx8nf_c3f5a989-6421-4bc7-a516-b37c825d8bbc/ovn-controller/0.log" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.631118 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" event={"ID":"c3f5a989-6421-4bc7-a516-b37c825d8bbc","Type":"ContainerDied","Data":"3c022f138769ab1d64f295dda0b5d503db70d70899c44ff8fa396d4de1d9824d"} Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.631246 4787 scope.go:117] "RemoveContainer" containerID="08b8f23afaf8a16b6613bf1e8f2bfea737d61a693f5a6f6c3c2213cf3202a093" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.631189 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-bx8nf" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.647486 4787 scope.go:117] "RemoveContainer" containerID="0a88782256d12e4f68943f17581c299dadb45610e50fa5f9d7d4abe8b2683d0c" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.680523 4787 scope.go:117] "RemoveContainer" containerID="cc5591e5019ecab1cd7f308fd4225a813788c4829f5f2116f87acae42723ab8f" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.738215 4787 scope.go:117] "RemoveContainer" containerID="24b3c01eaa0270ba72a07ce6d2bfd7ca5ecb6e9638678c17e4fcba06264572df" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.769578 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bx8nf"] Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.780450 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-bx8nf"] Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.800085 4787 scope.go:117] "RemoveContainer" containerID="43cc80d7f73f35a6cdb37d2ead98148e82faaabafd51b2f32b9eb2fce89c9516" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.822204 4787 scope.go:117] "RemoveContainer" containerID="b64a92a23d16e5d6b08b58c5ee0c2307ead64523a3e65fd1cf13771160c386d3" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.865238 4787 scope.go:117] "RemoveContainer" containerID="ef1775daa7cc7270a6327a6c2f1ae1033bb876b1be2cc7a1852d51675574b4d4" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.889963 4787 scope.go:117] "RemoveContainer" containerID="5f85e438a152351488b5e195ac0e5972ccf344604e0bd3ee7cc46731ea88c53e" Dec 03 17:25:10 crc kubenswrapper[4787]: I1203 17:25:10.917742 4787 scope.go:117] "RemoveContainer" containerID="988247b99f3f5aaea4c9ec9d094d204e6099b36e03f8695a33f43bb33bfe71f3" Dec 03 17:25:11 crc kubenswrapper[4787]: I1203 17:25:11.640105 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" event={"ID":"68802a81-1754-4f82-a98f-5f45fc7276d8","Type":"ContainerStarted","Data":"2a81d4c42880ab0b2fd4db4fa7e2f89b014af5b0f79d133d3f7164f937844c8a"} Dec 03 17:25:11 crc kubenswrapper[4787]: I1203 17:25:11.640402 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" event={"ID":"68802a81-1754-4f82-a98f-5f45fc7276d8","Type":"ContainerStarted","Data":"d95cafdfea0c3c4c305902743db0e1b02e3094219db5c003de6542ccd781ac45"} Dec 03 17:25:11 crc kubenswrapper[4787]: I1203 17:25:11.640415 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" event={"ID":"68802a81-1754-4f82-a98f-5f45fc7276d8","Type":"ContainerStarted","Data":"9f33f2f3d0a4c02826c3054d98b7ac6da778db46a44bf94533cb12f2a9f6fc8f"} Dec 03 17:25:11 crc kubenswrapper[4787]: I1203 17:25:11.640426 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" event={"ID":"68802a81-1754-4f82-a98f-5f45fc7276d8","Type":"ContainerStarted","Data":"5b2ff60d714f12ca4afa3bde03939e0f7a0b2a6aae5e93561d31797454e15deb"} Dec 03 17:25:11 crc kubenswrapper[4787]: I1203 17:25:11.640434 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" event={"ID":"68802a81-1754-4f82-a98f-5f45fc7276d8","Type":"ContainerStarted","Data":"9bc71ce2f558b087c4f44830e8dcdcca39a4fc518ee7a3ade2034bd1c4a6a6d1"} Dec 03 17:25:11 crc kubenswrapper[4787]: I1203 17:25:11.640446 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" event={"ID":"68802a81-1754-4f82-a98f-5f45fc7276d8","Type":"ContainerStarted","Data":"b132258810a1e2f174ed1805175c26b107cc1ee3398537624b2f574a45b3014a"} Dec 03 17:25:11 crc kubenswrapper[4787]: I1203 17:25:11.786788 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3f5a989-6421-4bc7-a516-b37c825d8bbc" path="/var/lib/kubelet/pods/c3f5a989-6421-4bc7-a516-b37c825d8bbc/volumes" Dec 03 17:25:14 crc kubenswrapper[4787]: I1203 17:25:14.656052 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" event={"ID":"68802a81-1754-4f82-a98f-5f45fc7276d8","Type":"ContainerStarted","Data":"6a235bdf08b3228c63ac8ddd4904b8f557595677b58abf1284afb7c7efdc6ec1"} Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.358572 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx"] Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.359388 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.362362 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-7pmhb" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.362473 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.363084 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.457784 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n95s\" (UniqueName: \"kubernetes.io/projected/d696b783-ca2e-495e-83fe-d33ab80fdf3f-kube-api-access-6n95s\") pod \"obo-prometheus-operator-668cf9dfbb-84bqx\" (UID: \"d696b783-ca2e-495e-83fe-d33ab80fdf3f\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.480396 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9"] Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.481091 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.483033 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-d7jsv" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.483248 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.488087 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9"] Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.488847 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.558531 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n95s\" (UniqueName: \"kubernetes.io/projected/d696b783-ca2e-495e-83fe-d33ab80fdf3f-kube-api-access-6n95s\") pod \"obo-prometheus-operator-668cf9dfbb-84bqx\" (UID: \"d696b783-ca2e-495e-83fe-d33ab80fdf3f\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.578268 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n95s\" (UniqueName: \"kubernetes.io/projected/d696b783-ca2e-495e-83fe-d33ab80fdf3f-kube-api-access-6n95s\") pod \"obo-prometheus-operator-668cf9dfbb-84bqx\" (UID: \"d696b783-ca2e-495e-83fe-d33ab80fdf3f\") " pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.659976 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ede219fc-7a66-484b-831f-6b242694494c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9\" (UID: \"ede219fc-7a66-484b-831f-6b242694494c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.660048 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ede219fc-7a66-484b-831f-6b242694494c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9\" (UID: \"ede219fc-7a66-484b-831f-6b242694494c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.660087 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1801fa2e-073e-486f-bd71-6abab103dd9f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9\" (UID: \"1801fa2e-073e-486f-bd71-6abab103dd9f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.660116 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1801fa2e-073e-486f-bd71-6abab103dd9f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9\" (UID: \"1801fa2e-073e-486f-bd71-6abab103dd9f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.675799 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.693001 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-ppms6"] Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.693811 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.696481 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-cxxzz" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.705342 4787 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators_d696b783-ca2e-495e-83fe-d33ab80fdf3f_0(52421c56db4b4406cfefb9875dd9de7531a24a9ab8c81b2a88b3499b461b9472): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.705404 4787 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators_d696b783-ca2e-495e-83fe-d33ab80fdf3f_0(52421c56db4b4406cfefb9875dd9de7531a24a9ab8c81b2a88b3499b461b9472): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.705424 4787 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators_d696b783-ca2e-495e-83fe-d33ab80fdf3f_0(52421c56db4b4406cfefb9875dd9de7531a24a9ab8c81b2a88b3499b461b9472): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.705463 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators(d696b783-ca2e-495e-83fe-d33ab80fdf3f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators(d696b783-ca2e-495e-83fe-d33ab80fdf3f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators_d696b783-ca2e-495e-83fe-d33ab80fdf3f_0(52421c56db4b4406cfefb9875dd9de7531a24a9ab8c81b2a88b3499b461b9472): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" podUID="d696b783-ca2e-495e-83fe-d33ab80fdf3f" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.705846 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.761673 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ede219fc-7a66-484b-831f-6b242694494c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9\" (UID: \"ede219fc-7a66-484b-831f-6b242694494c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.762049 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ede219fc-7a66-484b-831f-6b242694494c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9\" (UID: \"ede219fc-7a66-484b-831f-6b242694494c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.762081 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1801fa2e-073e-486f-bd71-6abab103dd9f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9\" (UID: \"1801fa2e-073e-486f-bd71-6abab103dd9f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.762120 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1801fa2e-073e-486f-bd71-6abab103dd9f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9\" (UID: \"1801fa2e-073e-486f-bd71-6abab103dd9f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.765568 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/ede219fc-7a66-484b-831f-6b242694494c-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9\" (UID: \"ede219fc-7a66-484b-831f-6b242694494c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.768574 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1801fa2e-073e-486f-bd71-6abab103dd9f-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9\" (UID: \"1801fa2e-073e-486f-bd71-6abab103dd9f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.772698 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ede219fc-7a66-484b-831f-6b242694494c-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9\" (UID: \"ede219fc-7a66-484b-831f-6b242694494c\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.773279 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1801fa2e-073e-486f-bd71-6abab103dd9f-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9\" (UID: \"1801fa2e-073e-486f-bd71-6abab103dd9f\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.799762 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.801622 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5446b9c989-6cfhr"] Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.802413 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.805784 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-zwjxb" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.807889 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.825824 4787 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators_ede219fc-7a66-484b-831f-6b242694494c_0(444a7fb22ec19335558001d83bfb5bda713bdf3fb6ffb326cf8a71c3fc94aaf0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.825903 4787 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators_ede219fc-7a66-484b-831f-6b242694494c_0(444a7fb22ec19335558001d83bfb5bda713bdf3fb6ffb326cf8a71c3fc94aaf0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.825937 4787 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators_ede219fc-7a66-484b-831f-6b242694494c_0(444a7fb22ec19335558001d83bfb5bda713bdf3fb6ffb326cf8a71c3fc94aaf0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.825993 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators(ede219fc-7a66-484b-831f-6b242694494c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators(ede219fc-7a66-484b-831f-6b242694494c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators_ede219fc-7a66-484b-831f-6b242694494c_0(444a7fb22ec19335558001d83bfb5bda713bdf3fb6ffb326cf8a71c3fc94aaf0): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" podUID="ede219fc-7a66-484b-831f-6b242694494c" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.835941 4787 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators_1801fa2e-073e-486f-bd71-6abab103dd9f_0(9bd291badb04475c404008f7a1419bce289f2b05f28cb0a46c7cdf9548dd6e8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.835992 4787 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators_1801fa2e-073e-486f-bd71-6abab103dd9f_0(9bd291badb04475c404008f7a1419bce289f2b05f28cb0a46c7cdf9548dd6e8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.836009 4787 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators_1801fa2e-073e-486f-bd71-6abab103dd9f_0(9bd291badb04475c404008f7a1419bce289f2b05f28cb0a46c7cdf9548dd6e8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:15 crc kubenswrapper[4787]: E1203 17:25:15.836063 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators(1801fa2e-073e-486f-bd71-6abab103dd9f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators(1801fa2e-073e-486f-bd71-6abab103dd9f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators_1801fa2e-073e-486f-bd71-6abab103dd9f_0(9bd291badb04475c404008f7a1419bce289f2b05f28cb0a46c7cdf9548dd6e8b): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" podUID="1801fa2e-073e-486f-bd71-6abab103dd9f" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.863589 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sjvv\" (UniqueName: \"kubernetes.io/projected/bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87-kube-api-access-6sjvv\") pod \"observability-operator-d8bb48f5d-ppms6\" (UID: \"bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87\") " pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.863687 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-ppms6\" (UID: \"bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87\") " pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.965060 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sjvv\" (UniqueName: \"kubernetes.io/projected/bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87-kube-api-access-6sjvv\") pod \"observability-operator-d8bb48f5d-ppms6\" (UID: \"bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87\") " pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.965128 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qxprw\" (UniqueName: \"kubernetes.io/projected/3651062d-0769-4da5-bb95-7c5987fc2b7e-kube-api-access-qxprw\") pod \"perses-operator-5446b9c989-6cfhr\" (UID: \"3651062d-0769-4da5-bb95-7c5987fc2b7e\") " pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.965176 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-ppms6\" (UID: \"bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87\") " pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.965199 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/3651062d-0769-4da5-bb95-7c5987fc2b7e-openshift-service-ca\") pod \"perses-operator-5446b9c989-6cfhr\" (UID: \"3651062d-0769-4da5-bb95-7c5987fc2b7e\") " pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.970351 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87-observability-operator-tls\") pod \"observability-operator-d8bb48f5d-ppms6\" (UID: \"bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87\") " pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:15 crc kubenswrapper[4787]: I1203 17:25:15.980596 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sjvv\" (UniqueName: \"kubernetes.io/projected/bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87-kube-api-access-6sjvv\") pod \"observability-operator-d8bb48f5d-ppms6\" (UID: \"bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87\") " pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:16 crc kubenswrapper[4787]: I1203 17:25:16.035411 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:16 crc kubenswrapper[4787]: E1203 17:25:16.065041 4787 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-ppms6_openshift-operators_bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87_0(adfde01878a897af44a5ee192ae40b2eae4595081fbc82996369b6960cfadefb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:25:16 crc kubenswrapper[4787]: E1203 17:25:16.065103 4787 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-ppms6_openshift-operators_bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87_0(adfde01878a897af44a5ee192ae40b2eae4595081fbc82996369b6960cfadefb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:16 crc kubenswrapper[4787]: E1203 17:25:16.065130 4787 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-ppms6_openshift-operators_bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87_0(adfde01878a897af44a5ee192ae40b2eae4595081fbc82996369b6960cfadefb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:16 crc kubenswrapper[4787]: E1203 17:25:16.065176 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-ppms6_openshift-operators(bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-ppms6_openshift-operators(bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-ppms6_openshift-operators_bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87_0(adfde01878a897af44a5ee192ae40b2eae4595081fbc82996369b6960cfadefb): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" podUID="bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87" Dec 03 17:25:16 crc kubenswrapper[4787]: I1203 17:25:16.066988 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qxprw\" (UniqueName: \"kubernetes.io/projected/3651062d-0769-4da5-bb95-7c5987fc2b7e-kube-api-access-qxprw\") pod \"perses-operator-5446b9c989-6cfhr\" (UID: \"3651062d-0769-4da5-bb95-7c5987fc2b7e\") " pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:16 crc kubenswrapper[4787]: I1203 17:25:16.067130 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/3651062d-0769-4da5-bb95-7c5987fc2b7e-openshift-service-ca\") pod \"perses-operator-5446b9c989-6cfhr\" (UID: \"3651062d-0769-4da5-bb95-7c5987fc2b7e\") " pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:16 crc kubenswrapper[4787]: I1203 17:25:16.070917 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/3651062d-0769-4da5-bb95-7c5987fc2b7e-openshift-service-ca\") pod \"perses-operator-5446b9c989-6cfhr\" (UID: \"3651062d-0769-4da5-bb95-7c5987fc2b7e\") " pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:16 crc kubenswrapper[4787]: I1203 17:25:16.084732 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qxprw\" (UniqueName: \"kubernetes.io/projected/3651062d-0769-4da5-bb95-7c5987fc2b7e-kube-api-access-qxprw\") pod \"perses-operator-5446b9c989-6cfhr\" (UID: \"3651062d-0769-4da5-bb95-7c5987fc2b7e\") " pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:16 crc kubenswrapper[4787]: I1203 17:25:16.133763 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:16 crc kubenswrapper[4787]: E1203 17:25:16.153236 4787 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-6cfhr_openshift-operators_3651062d-0769-4da5-bb95-7c5987fc2b7e_0(86c8fcf5c98dcca4f6f9c057faf563da39dd554c0e8628238f02c308131b58c8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:25:16 crc kubenswrapper[4787]: E1203 17:25:16.153303 4787 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-6cfhr_openshift-operators_3651062d-0769-4da5-bb95-7c5987fc2b7e_0(86c8fcf5c98dcca4f6f9c057faf563da39dd554c0e8628238f02c308131b58c8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:16 crc kubenswrapper[4787]: E1203 17:25:16.153330 4787 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-6cfhr_openshift-operators_3651062d-0769-4da5-bb95-7c5987fc2b7e_0(86c8fcf5c98dcca4f6f9c057faf563da39dd554c0e8628238f02c308131b58c8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:16 crc kubenswrapper[4787]: E1203 17:25:16.153386 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-6cfhr_openshift-operators(3651062d-0769-4da5-bb95-7c5987fc2b7e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-6cfhr_openshift-operators(3651062d-0769-4da5-bb95-7c5987fc2b7e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-6cfhr_openshift-operators_3651062d-0769-4da5-bb95-7c5987fc2b7e_0(86c8fcf5c98dcca4f6f9c057faf563da39dd554c0e8628238f02c308131b58c8): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" podUID="3651062d-0769-4da5-bb95-7c5987fc2b7e" Dec 03 17:25:16 crc kubenswrapper[4787]: I1203 17:25:16.668268 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" event={"ID":"68802a81-1754-4f82-a98f-5f45fc7276d8","Type":"ContainerStarted","Data":"99d7b5440fdda974cd4826f6266609af7b077a05ec3792ec4a12f9711c4c6154"} Dec 03 17:25:16 crc kubenswrapper[4787]: I1203 17:25:16.668732 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:16 crc kubenswrapper[4787]: I1203 17:25:16.704447 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:16 crc kubenswrapper[4787]: I1203 17:25:16.707579 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" podStartSLOduration=7.707567508 podStartE2EDuration="7.707567508s" podCreationTimestamp="2025-12-03 17:25:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:25:16.705100785 +0000 UTC m=+733.522571764" watchObservedRunningTime="2025-12-03 17:25:16.707567508 +0000 UTC m=+733.525038467" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.267378 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-ppms6"] Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.267488 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.267897 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.273889 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx"] Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.274002 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.274491 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.293516 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9"] Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.293605 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.293974 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.294845 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-6cfhr"] Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.294892 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.295109 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.316042 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9"] Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.316134 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.316492 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.327236 4787 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-ppms6_openshift-operators_bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87_0(535a34c0f360c69a787ffe92d4ae6aa29a53918c691fa61f8895fefdf34b3057): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.327301 4787 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-ppms6_openshift-operators_bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87_0(535a34c0f360c69a787ffe92d4ae6aa29a53918c691fa61f8895fefdf34b3057): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.327326 4787 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-ppms6_openshift-operators_bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87_0(535a34c0f360c69a787ffe92d4ae6aa29a53918c691fa61f8895fefdf34b3057): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.327364 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"observability-operator-d8bb48f5d-ppms6_openshift-operators(bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"observability-operator-d8bb48f5d-ppms6_openshift-operators(bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_observability-operator-d8bb48f5d-ppms6_openshift-operators_bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87_0(535a34c0f360c69a787ffe92d4ae6aa29a53918c691fa61f8895fefdf34b3057): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" podUID="bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.336232 4787 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators_d696b783-ca2e-495e-83fe-d33ab80fdf3f_0(4c4f91c41159fa3ceccf327590a34cf0dd591b10712d4da0f8c29d00d9f23472): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.336324 4787 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators_d696b783-ca2e-495e-83fe-d33ab80fdf3f_0(4c4f91c41159fa3ceccf327590a34cf0dd591b10712d4da0f8c29d00d9f23472): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.336347 4787 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators_d696b783-ca2e-495e-83fe-d33ab80fdf3f_0(4c4f91c41159fa3ceccf327590a34cf0dd591b10712d4da0f8c29d00d9f23472): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.336397 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators(d696b783-ca2e-495e-83fe-d33ab80fdf3f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators(d696b783-ca2e-495e-83fe-d33ab80fdf3f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-668cf9dfbb-84bqx_openshift-operators_d696b783-ca2e-495e-83fe-d33ab80fdf3f_0(4c4f91c41159fa3ceccf327590a34cf0dd591b10712d4da0f8c29d00d9f23472): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" podUID="d696b783-ca2e-495e-83fe-d33ab80fdf3f" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.345034 4787 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators_ede219fc-7a66-484b-831f-6b242694494c_0(3107215d00f559ec09dd6cc0aeadffe981b0e79f870184385dde881a7ff8a0ab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.345109 4787 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators_ede219fc-7a66-484b-831f-6b242694494c_0(3107215d00f559ec09dd6cc0aeadffe981b0e79f870184385dde881a7ff8a0ab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.345135 4787 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators_ede219fc-7a66-484b-831f-6b242694494c_0(3107215d00f559ec09dd6cc0aeadffe981b0e79f870184385dde881a7ff8a0ab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.345189 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators(ede219fc-7a66-484b-831f-6b242694494c)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators(ede219fc-7a66-484b-831f-6b242694494c)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_openshift-operators_ede219fc-7a66-484b-831f-6b242694494c_0(3107215d00f559ec09dd6cc0aeadffe981b0e79f870184385dde881a7ff8a0ab): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" podUID="ede219fc-7a66-484b-831f-6b242694494c" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.352039 4787 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-6cfhr_openshift-operators_3651062d-0769-4da5-bb95-7c5987fc2b7e_0(ae6a3f3326cd49aa0006ef0abc08ad1561d40f2fc540d6eefbeb24ff5ade3899): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.352122 4787 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-6cfhr_openshift-operators_3651062d-0769-4da5-bb95-7c5987fc2b7e_0(ae6a3f3326cd49aa0006ef0abc08ad1561d40f2fc540d6eefbeb24ff5ade3899): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.352148 4787 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-6cfhr_openshift-operators_3651062d-0769-4da5-bb95-7c5987fc2b7e_0(ae6a3f3326cd49aa0006ef0abc08ad1561d40f2fc540d6eefbeb24ff5ade3899): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.352205 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"perses-operator-5446b9c989-6cfhr_openshift-operators(3651062d-0769-4da5-bb95-7c5987fc2b7e)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"perses-operator-5446b9c989-6cfhr_openshift-operators(3651062d-0769-4da5-bb95-7c5987fc2b7e)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_perses-operator-5446b9c989-6cfhr_openshift-operators_3651062d-0769-4da5-bb95-7c5987fc2b7e_0(ae6a3f3326cd49aa0006ef0abc08ad1561d40f2fc540d6eefbeb24ff5ade3899): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" podUID="3651062d-0769-4da5-bb95-7c5987fc2b7e" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.378857 4787 log.go:32] "RunPodSandbox from runtime service failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators_1801fa2e-073e-486f-bd71-6abab103dd9f_0(95e6d1b6a28fe7024ce7512fcd75bbfab68e0d70e87ee874f6a36839941a4c60): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.378940 4787 kuberuntime_sandbox.go:72] "Failed to create sandbox for pod" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators_1801fa2e-073e-486f-bd71-6abab103dd9f_0(95e6d1b6a28fe7024ce7512fcd75bbfab68e0d70e87ee874f6a36839941a4c60): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.378966 4787 kuberuntime_manager.go:1170] "CreatePodSandbox for pod failed" err="rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators_1801fa2e-073e-486f-bd71-6abab103dd9f_0(95e6d1b6a28fe7024ce7512fcd75bbfab68e0d70e87ee874f6a36839941a4c60): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:17 crc kubenswrapper[4787]: E1203 17:25:17.379041 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"CreatePodSandbox\" for \"obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators(1801fa2e-073e-486f-bd71-6abab103dd9f)\" with CreatePodSandboxError: \"Failed to create sandbox for pod \\\"obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators(1801fa2e-073e-486f-bd71-6abab103dd9f)\\\": rpc error: code = Unknown desc = failed to create pod network sandbox k8s_obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_openshift-operators_1801fa2e-073e-486f-bd71-6abab103dd9f_0(95e6d1b6a28fe7024ce7512fcd75bbfab68e0d70e87ee874f6a36839941a4c60): no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\"" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" podUID="1801fa2e-073e-486f-bd71-6abab103dd9f" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.674459 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.674505 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:17 crc kubenswrapper[4787]: I1203 17:25:17.715749 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:21 crc kubenswrapper[4787]: I1203 17:25:21.765949 4787 scope.go:117] "RemoveContainer" containerID="d24796ebc7e29b1d6ddb686edefd44466c78b350cb28a0f36ea374203323649a" Dec 03 17:25:22 crc kubenswrapper[4787]: I1203 17:25:22.699224 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-xcz4f_bd5617c0-aa4f-4f21-a131-15af831725ec/kube-multus/2.log" Dec 03 17:25:22 crc kubenswrapper[4787]: I1203 17:25:22.699827 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-xcz4f" event={"ID":"bd5617c0-aa4f-4f21-a131-15af831725ec","Type":"ContainerStarted","Data":"badf064c8d757d562afdf8463b02b42b40cc4902ec30e4d79169387b1cfb1419"} Dec 03 17:25:28 crc kubenswrapper[4787]: I1203 17:25:28.765892 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:28 crc kubenswrapper[4787]: I1203 17:25:28.766993 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:28 crc kubenswrapper[4787]: I1203 17:25:28.997238 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5446b9c989-6cfhr"] Dec 03 17:25:29 crc kubenswrapper[4787]: I1203 17:25:29.739638 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" event={"ID":"3651062d-0769-4da5-bb95-7c5987fc2b7e","Type":"ContainerStarted","Data":"47bb4c1217cef05db8ce3c1d42edea1f00a28e0e8d46281fbbdbd8b0e2e2feba"} Dec 03 17:25:31 crc kubenswrapper[4787]: I1203 17:25:31.765457 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:31 crc kubenswrapper[4787]: I1203 17:25:31.765666 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:31 crc kubenswrapper[4787]: I1203 17:25:31.765671 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:31 crc kubenswrapper[4787]: I1203 17:25:31.766325 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" Dec 03 17:25:31 crc kubenswrapper[4787]: I1203 17:25:31.766400 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" Dec 03 17:25:31 crc kubenswrapper[4787]: I1203 17:25:31.766680 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" Dec 03 17:25:32 crc kubenswrapper[4787]: I1203 17:25:32.130555 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9"] Dec 03 17:25:32 crc kubenswrapper[4787]: I1203 17:25:32.199130 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9"] Dec 03 17:25:32 crc kubenswrapper[4787]: W1203 17:25:32.207100 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podede219fc_7a66_484b_831f_6b242694494c.slice/crio-7d1b1f98796a0e7e6d5bee87d54879866961406712213610f00079849bc3aee3 WatchSource:0}: Error finding container 7d1b1f98796a0e7e6d5bee87d54879866961406712213610f00079849bc3aee3: Status 404 returned error can't find the container with id 7d1b1f98796a0e7e6d5bee87d54879866961406712213610f00079849bc3aee3 Dec 03 17:25:32 crc kubenswrapper[4787]: I1203 17:25:32.444037 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx"] Dec 03 17:25:32 crc kubenswrapper[4787]: W1203 17:25:32.449978 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd696b783_ca2e_495e_83fe_d33ab80fdf3f.slice/crio-d8df16df7f28a819ff527598e90a736d291e4799c7ab0dd8a4c971c6687af3ed WatchSource:0}: Error finding container d8df16df7f28a819ff527598e90a736d291e4799c7ab0dd8a4c971c6687af3ed: Status 404 returned error can't find the container with id d8df16df7f28a819ff527598e90a736d291e4799c7ab0dd8a4c971c6687af3ed Dec 03 17:25:32 crc kubenswrapper[4787]: I1203 17:25:32.765046 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:32 crc kubenswrapper[4787]: I1203 17:25:32.765609 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:32 crc kubenswrapper[4787]: I1203 17:25:32.773579 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" event={"ID":"1801fa2e-073e-486f-bd71-6abab103dd9f","Type":"ContainerStarted","Data":"25cf04eb9824b7cfaf1fdc389803940a21b31aa7ed49fb120692541dfc740001"} Dec 03 17:25:32 crc kubenswrapper[4787]: I1203 17:25:32.775866 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" event={"ID":"d696b783-ca2e-495e-83fe-d33ab80fdf3f","Type":"ContainerStarted","Data":"d8df16df7f28a819ff527598e90a736d291e4799c7ab0dd8a4c971c6687af3ed"} Dec 03 17:25:32 crc kubenswrapper[4787]: I1203 17:25:32.777370 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" event={"ID":"ede219fc-7a66-484b-831f-6b242694494c","Type":"ContainerStarted","Data":"7d1b1f98796a0e7e6d5bee87d54879866961406712213610f00079849bc3aee3"} Dec 03 17:25:33 crc kubenswrapper[4787]: I1203 17:25:33.025514 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-d8bb48f5d-ppms6"] Dec 03 17:25:33 crc kubenswrapper[4787]: W1203 17:25:33.033224 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf31c2a1_79e3_47bd_8e8c_f9c6ce432f87.slice/crio-b541320340a1542eda8d0993ea7c09563b8003bae54b76833a18097001ee4cf9 WatchSource:0}: Error finding container b541320340a1542eda8d0993ea7c09563b8003bae54b76833a18097001ee4cf9: Status 404 returned error can't find the container with id b541320340a1542eda8d0993ea7c09563b8003bae54b76833a18097001ee4cf9 Dec 03 17:25:33 crc kubenswrapper[4787]: I1203 17:25:33.786635 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" event={"ID":"bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87","Type":"ContainerStarted","Data":"b541320340a1542eda8d0993ea7c09563b8003bae54b76833a18097001ee4cf9"} Dec 03 17:25:40 crc kubenswrapper[4787]: I1203 17:25:40.057401 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-swdrj" Dec 03 17:25:41 crc kubenswrapper[4787]: I1203 17:25:41.860557 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" event={"ID":"d696b783-ca2e-495e-83fe-d33ab80fdf3f","Type":"ContainerStarted","Data":"c2b754c8e7d8ad4e5862954740dd947fe9aadfe4027cd62a4eb9e25b4fc95c4d"} Dec 03 17:25:41 crc kubenswrapper[4787]: I1203 17:25:41.863572 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" event={"ID":"ede219fc-7a66-484b-831f-6b242694494c","Type":"ContainerStarted","Data":"0e6618c1e42abafa930a1e58cd55db3335dc873a7c5c7fbc70bb6348fdc3700f"} Dec 03 17:25:41 crc kubenswrapper[4787]: I1203 17:25:41.865590 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" event={"ID":"1801fa2e-073e-486f-bd71-6abab103dd9f","Type":"ContainerStarted","Data":"342fe57ebaba7d25f9e3d9209c0ac12481d85fd9e6bfb6bde62cd749c05c7970"} Dec 03 17:25:41 crc kubenswrapper[4787]: I1203 17:25:41.867297 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" event={"ID":"3651062d-0769-4da5-bb95-7c5987fc2b7e","Type":"ContainerStarted","Data":"6d62c02370ca21b1ed4c69deea3bdee1438e4fe01abb74e7da9e99e82abacd0b"} Dec 03 17:25:41 crc kubenswrapper[4787]: I1203 17:25:41.867779 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:41 crc kubenswrapper[4787]: I1203 17:25:41.880548 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-668cf9dfbb-84bqx" podStartSLOduration=18.163711859 podStartE2EDuration="26.880527695s" podCreationTimestamp="2025-12-03 17:25:15 +0000 UTC" firstStartedPulling="2025-12-03 17:25:32.45318383 +0000 UTC m=+749.270654789" lastFinishedPulling="2025-12-03 17:25:41.169999666 +0000 UTC m=+757.987470625" observedRunningTime="2025-12-03 17:25:41.878529614 +0000 UTC m=+758.696000593" watchObservedRunningTime="2025-12-03 17:25:41.880527695 +0000 UTC m=+758.697998654" Dec 03 17:25:41 crc kubenswrapper[4787]: I1203 17:25:41.944943 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" podStartSLOduration=14.865849702 podStartE2EDuration="26.94492754s" podCreationTimestamp="2025-12-03 17:25:15 +0000 UTC" firstStartedPulling="2025-12-03 17:25:29.00252912 +0000 UTC m=+745.820000079" lastFinishedPulling="2025-12-03 17:25:41.081606958 +0000 UTC m=+757.899077917" observedRunningTime="2025-12-03 17:25:41.941949904 +0000 UTC m=+758.759420863" watchObservedRunningTime="2025-12-03 17:25:41.94492754 +0000 UTC m=+758.762398499" Dec 03 17:25:41 crc kubenswrapper[4787]: I1203 17:25:41.948255 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9" podStartSLOduration=17.981319571 podStartE2EDuration="26.948244025s" podCreationTimestamp="2025-12-03 17:25:15 +0000 UTC" firstStartedPulling="2025-12-03 17:25:32.168635092 +0000 UTC m=+748.986106051" lastFinishedPulling="2025-12-03 17:25:41.135559546 +0000 UTC m=+757.953030505" observedRunningTime="2025-12-03 17:25:41.906762615 +0000 UTC m=+758.724233574" watchObservedRunningTime="2025-12-03 17:25:41.948244025 +0000 UTC m=+758.765714984" Dec 03 17:25:41 crc kubenswrapper[4787]: I1203 17:25:41.960823 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9" podStartSLOduration=18.093435914 podStartE2EDuration="26.960802815s" podCreationTimestamp="2025-12-03 17:25:15 +0000 UTC" firstStartedPulling="2025-12-03 17:25:32.209147197 +0000 UTC m=+749.026618156" lastFinishedPulling="2025-12-03 17:25:41.076514098 +0000 UTC m=+757.893985057" observedRunningTime="2025-12-03 17:25:41.955746926 +0000 UTC m=+758.773217895" watchObservedRunningTime="2025-12-03 17:25:41.960802815 +0000 UTC m=+758.778273774" Dec 03 17:25:43 crc kubenswrapper[4787]: I1203 17:25:43.881291 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" event={"ID":"bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87","Type":"ContainerStarted","Data":"d0537ceed31b8ac12f0c46f1b88f7412e8fecdf9476e2a5dedc74024b23783af"} Dec 03 17:25:43 crc kubenswrapper[4787]: I1203 17:25:43.883095 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:43 crc kubenswrapper[4787]: I1203 17:25:43.912239 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" podStartSLOduration=18.559809207 podStartE2EDuration="28.912219542s" podCreationTimestamp="2025-12-03 17:25:15 +0000 UTC" firstStartedPulling="2025-12-03 17:25:33.035214247 +0000 UTC m=+749.852685206" lastFinishedPulling="2025-12-03 17:25:43.387624582 +0000 UTC m=+760.205095541" observedRunningTime="2025-12-03 17:25:43.909935033 +0000 UTC m=+760.727406032" watchObservedRunningTime="2025-12-03 17:25:43.912219542 +0000 UTC m=+760.729690501" Dec 03 17:25:43 crc kubenswrapper[4787]: I1203 17:25:43.921069 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-d8bb48f5d-ppms6" Dec 03 17:25:46 crc kubenswrapper[4787]: I1203 17:25:46.137529 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" Dec 03 17:25:46 crc kubenswrapper[4787]: I1203 17:25:46.792488 4787 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.473346 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-r8mcm"] Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.474968 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-r8mcm" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.478151 4787 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-frqc2" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.479189 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.481499 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.483580 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hrd48"] Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.484694 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-hrd48" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.488197 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-r8mcm"] Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.494476 4787 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-r4qt8" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.500643 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bx2z8\" (UniqueName: \"kubernetes.io/projected/249fd549-e9f6-4a50-9b84-b96e46f2791f-kube-api-access-bx2z8\") pod \"cert-manager-5b446d88c5-hrd48\" (UID: \"249fd549-e9f6-4a50-9b84-b96e46f2791f\") " pod="cert-manager/cert-manager-5b446d88c5-hrd48" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.500723 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl52z\" (UniqueName: \"kubernetes.io/projected/c0bb5afc-e67f-42c5-849e-2892c2bf7d3a-kube-api-access-gl52z\") pod \"cert-manager-cainjector-7f985d654d-r8mcm\" (UID: \"c0bb5afc-e67f-42c5-849e-2892c2bf7d3a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-r8mcm" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.504977 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hrd48"] Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.516571 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-f46ld"] Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.517344 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-f46ld" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.528533 4787 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-dh6b9" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.528942 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-f46ld"] Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.601435 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl52z\" (UniqueName: \"kubernetes.io/projected/c0bb5afc-e67f-42c5-849e-2892c2bf7d3a-kube-api-access-gl52z\") pod \"cert-manager-cainjector-7f985d654d-r8mcm\" (UID: \"c0bb5afc-e67f-42c5-849e-2892c2bf7d3a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-r8mcm" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.601493 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kfmx6\" (UniqueName: \"kubernetes.io/projected/f5b1552f-1b75-4000-b31e-4d247ea51a65-kube-api-access-kfmx6\") pod \"cert-manager-webhook-5655c58dd6-f46ld\" (UID: \"f5b1552f-1b75-4000-b31e-4d247ea51a65\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-f46ld" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.601555 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bx2z8\" (UniqueName: \"kubernetes.io/projected/249fd549-e9f6-4a50-9b84-b96e46f2791f-kube-api-access-bx2z8\") pod \"cert-manager-5b446d88c5-hrd48\" (UID: \"249fd549-e9f6-4a50-9b84-b96e46f2791f\") " pod="cert-manager/cert-manager-5b446d88c5-hrd48" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.638219 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bx2z8\" (UniqueName: \"kubernetes.io/projected/249fd549-e9f6-4a50-9b84-b96e46f2791f-kube-api-access-bx2z8\") pod \"cert-manager-5b446d88c5-hrd48\" (UID: \"249fd549-e9f6-4a50-9b84-b96e46f2791f\") " pod="cert-manager/cert-manager-5b446d88c5-hrd48" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.641430 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl52z\" (UniqueName: \"kubernetes.io/projected/c0bb5afc-e67f-42c5-849e-2892c2bf7d3a-kube-api-access-gl52z\") pod \"cert-manager-cainjector-7f985d654d-r8mcm\" (UID: \"c0bb5afc-e67f-42c5-849e-2892c2bf7d3a\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-r8mcm" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.702434 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kfmx6\" (UniqueName: \"kubernetes.io/projected/f5b1552f-1b75-4000-b31e-4d247ea51a65-kube-api-access-kfmx6\") pod \"cert-manager-webhook-5655c58dd6-f46ld\" (UID: \"f5b1552f-1b75-4000-b31e-4d247ea51a65\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-f46ld" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.728666 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kfmx6\" (UniqueName: \"kubernetes.io/projected/f5b1552f-1b75-4000-b31e-4d247ea51a65-kube-api-access-kfmx6\") pod \"cert-manager-webhook-5655c58dd6-f46ld\" (UID: \"f5b1552f-1b75-4000-b31e-4d247ea51a65\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-f46ld" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.796086 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-r8mcm" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.802192 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-hrd48" Dec 03 17:25:51 crc kubenswrapper[4787]: I1203 17:25:51.829353 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-f46ld" Dec 03 17:25:52 crc kubenswrapper[4787]: I1203 17:25:52.038398 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-f46ld"] Dec 03 17:25:52 crc kubenswrapper[4787]: W1203 17:25:52.050533 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5b1552f_1b75_4000_b31e_4d247ea51a65.slice/crio-5652b1173e4999bd7c1ec336cf7c67bdcabce4b090ca599bfa600270df5def86 WatchSource:0}: Error finding container 5652b1173e4999bd7c1ec336cf7c67bdcabce4b090ca599bfa600270df5def86: Status 404 returned error can't find the container with id 5652b1173e4999bd7c1ec336cf7c67bdcabce4b090ca599bfa600270df5def86 Dec 03 17:25:52 crc kubenswrapper[4787]: I1203 17:25:52.070888 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-hrd48"] Dec 03 17:25:52 crc kubenswrapper[4787]: W1203 17:25:52.074337 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod249fd549_e9f6_4a50_9b84_b96e46f2791f.slice/crio-d74cdd297e702f3aa5c3940f33586ba33ba79977ec76693f1c1eaa63fedc1c26 WatchSource:0}: Error finding container d74cdd297e702f3aa5c3940f33586ba33ba79977ec76693f1c1eaa63fedc1c26: Status 404 returned error can't find the container with id d74cdd297e702f3aa5c3940f33586ba33ba79977ec76693f1c1eaa63fedc1c26 Dec 03 17:25:52 crc kubenswrapper[4787]: I1203 17:25:52.355629 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-r8mcm"] Dec 03 17:25:52 crc kubenswrapper[4787]: W1203 17:25:52.362149 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc0bb5afc_e67f_42c5_849e_2892c2bf7d3a.slice/crio-c227ed8d1f90e8337337820e3bb956aced98b869d0419101bda830388cf1c0d3 WatchSource:0}: Error finding container c227ed8d1f90e8337337820e3bb956aced98b869d0419101bda830388cf1c0d3: Status 404 returned error can't find the container with id c227ed8d1f90e8337337820e3bb956aced98b869d0419101bda830388cf1c0d3 Dec 03 17:25:52 crc kubenswrapper[4787]: I1203 17:25:52.933222 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-f46ld" event={"ID":"f5b1552f-1b75-4000-b31e-4d247ea51a65","Type":"ContainerStarted","Data":"5652b1173e4999bd7c1ec336cf7c67bdcabce4b090ca599bfa600270df5def86"} Dec 03 17:25:52 crc kubenswrapper[4787]: I1203 17:25:52.934192 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-hrd48" event={"ID":"249fd549-e9f6-4a50-9b84-b96e46f2791f","Type":"ContainerStarted","Data":"d74cdd297e702f3aa5c3940f33586ba33ba79977ec76693f1c1eaa63fedc1c26"} Dec 03 17:25:52 crc kubenswrapper[4787]: I1203 17:25:52.935544 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-r8mcm" event={"ID":"c0bb5afc-e67f-42c5-849e-2892c2bf7d3a","Type":"ContainerStarted","Data":"c227ed8d1f90e8337337820e3bb956aced98b869d0419101bda830388cf1c0d3"} Dec 03 17:25:56 crc kubenswrapper[4787]: I1203 17:25:56.964223 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-r8mcm" event={"ID":"c0bb5afc-e67f-42c5-849e-2892c2bf7d3a","Type":"ContainerStarted","Data":"51d82205d7fa3c1537f81b5d35c7b5f30ac5c8dd72222af63252415480354176"} Dec 03 17:25:56 crc kubenswrapper[4787]: I1203 17:25:56.967556 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-f46ld" event={"ID":"f5b1552f-1b75-4000-b31e-4d247ea51a65","Type":"ContainerStarted","Data":"dbb50c150a6499edc39ccef61e24abdadf06559574279d71d81c3fc7d2e0dceb"} Dec 03 17:25:56 crc kubenswrapper[4787]: I1203 17:25:56.967991 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-f46ld" Dec 03 17:25:56 crc kubenswrapper[4787]: I1203 17:25:56.969799 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-hrd48" event={"ID":"249fd549-e9f6-4a50-9b84-b96e46f2791f","Type":"ContainerStarted","Data":"8cabdd25f4ffb954ceabcdeb7c2a2c2bcf25df501c428d8f7c8495f34f2a3762"} Dec 03 17:25:56 crc kubenswrapper[4787]: I1203 17:25:56.985434 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-r8mcm" podStartSLOduration=2.5214329920000003 podStartE2EDuration="5.985415382s" podCreationTimestamp="2025-12-03 17:25:51 +0000 UTC" firstStartedPulling="2025-12-03 17:25:52.36414219 +0000 UTC m=+769.181613149" lastFinishedPulling="2025-12-03 17:25:55.82812457 +0000 UTC m=+772.645595539" observedRunningTime="2025-12-03 17:25:56.984073447 +0000 UTC m=+773.801544416" watchObservedRunningTime="2025-12-03 17:25:56.985415382 +0000 UTC m=+773.802886341" Dec 03 17:25:57 crc kubenswrapper[4787]: I1203 17:25:57.004708 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-f46ld" podStartSLOduration=2.166417972 podStartE2EDuration="6.004689474s" podCreationTimestamp="2025-12-03 17:25:51 +0000 UTC" firstStartedPulling="2025-12-03 17:25:52.056254895 +0000 UTC m=+768.873725854" lastFinishedPulling="2025-12-03 17:25:55.894526397 +0000 UTC m=+772.711997356" observedRunningTime="2025-12-03 17:25:57.00100915 +0000 UTC m=+773.818480139" watchObservedRunningTime="2025-12-03 17:25:57.004689474 +0000 UTC m=+773.822160433" Dec 03 17:25:57 crc kubenswrapper[4787]: I1203 17:25:57.023719 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-hrd48" podStartSLOduration=2.349567571 podStartE2EDuration="6.02370158s" podCreationTimestamp="2025-12-03 17:25:51 +0000 UTC" firstStartedPulling="2025-12-03 17:25:52.077564629 +0000 UTC m=+768.895035588" lastFinishedPulling="2025-12-03 17:25:55.751698628 +0000 UTC m=+772.569169597" observedRunningTime="2025-12-03 17:25:57.021977306 +0000 UTC m=+773.839448285" watchObservedRunningTime="2025-12-03 17:25:57.02370158 +0000 UTC m=+773.841172539" Dec 03 17:26:01 crc kubenswrapper[4787]: I1203 17:26:01.832670 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-f46ld" Dec 03 17:26:18 crc kubenswrapper[4787]: I1203 17:26:18.989725 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:26:18 crc kubenswrapper[4787]: I1203 17:26:18.990455 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:26:28 crc kubenswrapper[4787]: I1203 17:26:28.953017 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j"] Dec 03 17:26:28 crc kubenswrapper[4787]: I1203 17:26:28.954528 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:28 crc kubenswrapper[4787]: I1203 17:26:28.956748 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 17:26:28 crc kubenswrapper[4787]: I1203 17:26:28.968000 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j"] Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.026831 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.026919 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vhw5\" (UniqueName: \"kubernetes.io/projected/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-kube-api-access-9vhw5\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.026984 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.128423 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.128513 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vhw5\" (UniqueName: \"kubernetes.io/projected/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-kube-api-access-9vhw5\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.128551 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.129112 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-util\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.129201 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-bundle\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.164551 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vhw5\" (UniqueName: \"kubernetes.io/projected/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-kube-api-access-9vhw5\") pod \"a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.275404 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.375016 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j"] Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.377414 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.393954 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j"] Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.434448 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2c8c\" (UniqueName: \"kubernetes.io/projected/74207f9d-ab04-4834-9813-5d89fe068d43-kube-api-access-n2c8c\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.434511 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.434688 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.537953 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2c8c\" (UniqueName: \"kubernetes.io/projected/74207f9d-ab04-4834-9813-5d89fe068d43-kube-api-access-n2c8c\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.538047 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.538178 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.538960 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-util\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.538982 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-bundle\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.556581 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2c8c\" (UniqueName: \"kubernetes.io/projected/74207f9d-ab04-4834-9813-5d89fe068d43-kube-api-access-n2c8c\") pod \"4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.727626 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.729585 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j"] Dec 03 17:26:29 crc kubenswrapper[4787]: I1203 17:26:29.958294 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j"] Dec 03 17:26:30 crc kubenswrapper[4787]: I1203 17:26:30.197976 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" event={"ID":"74207f9d-ab04-4834-9813-5d89fe068d43","Type":"ContainerStarted","Data":"ba1580dd32a33af09d2cc252b858e4db54a9ac0bf29e547645eaeeda7300c624"} Dec 03 17:26:30 crc kubenswrapper[4787]: I1203 17:26:30.198562 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" event={"ID":"74207f9d-ab04-4834-9813-5d89fe068d43","Type":"ContainerStarted","Data":"0c4322a63d9089d8c2e4c6c8860275371cbd09a11bfa72a23b2a08f3aae8d3de"} Dec 03 17:26:30 crc kubenswrapper[4787]: I1203 17:26:30.201993 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" event={"ID":"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0","Type":"ContainerStarted","Data":"443f071ccf331328464cb54ea245d00f485968049a1abe9ce989b851c2a31fb0"} Dec 03 17:26:30 crc kubenswrapper[4787]: I1203 17:26:30.202081 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" event={"ID":"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0","Type":"ContainerStarted","Data":"51a0d6b980ddc846819ccf04a73ee18805b1a0be42bca177c9f6c7d88ecef6f2"} Dec 03 17:26:31 crc kubenswrapper[4787]: I1203 17:26:31.208724 4787 generic.go:334] "Generic (PLEG): container finished" podID="74207f9d-ab04-4834-9813-5d89fe068d43" containerID="ba1580dd32a33af09d2cc252b858e4db54a9ac0bf29e547645eaeeda7300c624" exitCode=0 Dec 03 17:26:31 crc kubenswrapper[4787]: I1203 17:26:31.208808 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" event={"ID":"74207f9d-ab04-4834-9813-5d89fe068d43","Type":"ContainerDied","Data":"ba1580dd32a33af09d2cc252b858e4db54a9ac0bf29e547645eaeeda7300c624"} Dec 03 17:26:31 crc kubenswrapper[4787]: I1203 17:26:31.210726 4787 generic.go:334] "Generic (PLEG): container finished" podID="9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" containerID="443f071ccf331328464cb54ea245d00f485968049a1abe9ce989b851c2a31fb0" exitCode=0 Dec 03 17:26:31 crc kubenswrapper[4787]: I1203 17:26:31.210772 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" event={"ID":"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0","Type":"ContainerDied","Data":"443f071ccf331328464cb54ea245d00f485968049a1abe9ce989b851c2a31fb0"} Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.701941 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k9kw4"] Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.703873 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.711847 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k9kw4"] Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.774773 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-catalog-content\") pod \"redhat-operators-k9kw4\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.774853 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-utilities\") pod \"redhat-operators-k9kw4\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.774885 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6c2p\" (UniqueName: \"kubernetes.io/projected/d075806c-3c60-4139-8ecf-9c9a384d1c24-kube-api-access-b6c2p\") pod \"redhat-operators-k9kw4\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.876978 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-catalog-content\") pod \"redhat-operators-k9kw4\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.877057 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-utilities\") pod \"redhat-operators-k9kw4\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.877090 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6c2p\" (UniqueName: \"kubernetes.io/projected/d075806c-3c60-4139-8ecf-9c9a384d1c24-kube-api-access-b6c2p\") pod \"redhat-operators-k9kw4\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.877652 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-catalog-content\") pod \"redhat-operators-k9kw4\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.877775 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-utilities\") pod \"redhat-operators-k9kw4\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:32 crc kubenswrapper[4787]: I1203 17:26:32.902769 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6c2p\" (UniqueName: \"kubernetes.io/projected/d075806c-3c60-4139-8ecf-9c9a384d1c24-kube-api-access-b6c2p\") pod \"redhat-operators-k9kw4\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:33 crc kubenswrapper[4787]: I1203 17:26:33.091489 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:33 crc kubenswrapper[4787]: I1203 17:26:33.231900 4787 generic.go:334] "Generic (PLEG): container finished" podID="74207f9d-ab04-4834-9813-5d89fe068d43" containerID="e200d26905485691ae4819d0c1616021de8b4ff87262d4add07d7604cb407475" exitCode=0 Dec 03 17:26:33 crc kubenswrapper[4787]: I1203 17:26:33.232161 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" event={"ID":"74207f9d-ab04-4834-9813-5d89fe068d43","Type":"ContainerDied","Data":"e200d26905485691ae4819d0c1616021de8b4ff87262d4add07d7604cb407475"} Dec 03 17:26:33 crc kubenswrapper[4787]: I1203 17:26:33.237009 4787 generic.go:334] "Generic (PLEG): container finished" podID="9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" containerID="d1d345b6310066300f37293c4ea69781225c77da9bdedc30aef17f2d0f51a0b0" exitCode=0 Dec 03 17:26:33 crc kubenswrapper[4787]: I1203 17:26:33.237092 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" event={"ID":"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0","Type":"ContainerDied","Data":"d1d345b6310066300f37293c4ea69781225c77da9bdedc30aef17f2d0f51a0b0"} Dec 03 17:26:33 crc kubenswrapper[4787]: I1203 17:26:33.592767 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k9kw4"] Dec 03 17:26:34 crc kubenswrapper[4787]: I1203 17:26:34.244359 4787 generic.go:334] "Generic (PLEG): container finished" podID="74207f9d-ab04-4834-9813-5d89fe068d43" containerID="bb05b58d101680c72bfd9588754ac0fc148531af729c42bcdff52e27a1f33d5b" exitCode=0 Dec 03 17:26:34 crc kubenswrapper[4787]: I1203 17:26:34.244424 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" event={"ID":"74207f9d-ab04-4834-9813-5d89fe068d43","Type":"ContainerDied","Data":"bb05b58d101680c72bfd9588754ac0fc148531af729c42bcdff52e27a1f33d5b"} Dec 03 17:26:34 crc kubenswrapper[4787]: I1203 17:26:34.246129 4787 generic.go:334] "Generic (PLEG): container finished" podID="9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" containerID="54960c9b98edd613b4cd20c76a7265725085c5e66e0f66849ac301ffb337abb5" exitCode=0 Dec 03 17:26:34 crc kubenswrapper[4787]: I1203 17:26:34.246219 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" event={"ID":"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0","Type":"ContainerDied","Data":"54960c9b98edd613b4cd20c76a7265725085c5e66e0f66849ac301ffb337abb5"} Dec 03 17:26:34 crc kubenswrapper[4787]: I1203 17:26:34.247924 4787 generic.go:334] "Generic (PLEG): container finished" podID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerID="de9ec3f2fc3b8cb05db0812ee10c5f1f9559ee31403d271ae968254b655ac3bb" exitCode=0 Dec 03 17:26:34 crc kubenswrapper[4787]: I1203 17:26:34.247990 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k9kw4" event={"ID":"d075806c-3c60-4139-8ecf-9c9a384d1c24","Type":"ContainerDied","Data":"de9ec3f2fc3b8cb05db0812ee10c5f1f9559ee31403d271ae968254b655ac3bb"} Dec 03 17:26:34 crc kubenswrapper[4787]: I1203 17:26:34.248122 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k9kw4" event={"ID":"d075806c-3c60-4139-8ecf-9c9a384d1c24","Type":"ContainerStarted","Data":"3ab1892ec91374654437d1859136b601a7e5d5f9d0d50f006bc73411bf8d951c"} Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.255433 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k9kw4" event={"ID":"d075806c-3c60-4139-8ecf-9c9a384d1c24","Type":"ContainerStarted","Data":"5fb5197a75036fd07fe1267a8e3f9b3739a483f7c55b1bd48a635bf808c22715"} Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.538180 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.612345 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.716895 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-bundle\") pod \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.716962 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2c8c\" (UniqueName: \"kubernetes.io/projected/74207f9d-ab04-4834-9813-5d89fe068d43-kube-api-access-n2c8c\") pod \"74207f9d-ab04-4834-9813-5d89fe068d43\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.717015 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-util\") pod \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.717099 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9vhw5\" (UniqueName: \"kubernetes.io/projected/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-kube-api-access-9vhw5\") pod \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\" (UID: \"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0\") " Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.717151 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-bundle\") pod \"74207f9d-ab04-4834-9813-5d89fe068d43\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.717222 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-util\") pod \"74207f9d-ab04-4834-9813-5d89fe068d43\" (UID: \"74207f9d-ab04-4834-9813-5d89fe068d43\") " Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.718360 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-bundle" (OuterVolumeSpecName: "bundle") pod "74207f9d-ab04-4834-9813-5d89fe068d43" (UID: "74207f9d-ab04-4834-9813-5d89fe068d43"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.718557 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-bundle" (OuterVolumeSpecName: "bundle") pod "9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" (UID: "9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.721466 4787 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.721492 4787 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.728259 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74207f9d-ab04-4834-9813-5d89fe068d43-kube-api-access-n2c8c" (OuterVolumeSpecName: "kube-api-access-n2c8c") pod "74207f9d-ab04-4834-9813-5d89fe068d43" (UID: "74207f9d-ab04-4834-9813-5d89fe068d43"). InnerVolumeSpecName "kube-api-access-n2c8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.739288 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-kube-api-access-9vhw5" (OuterVolumeSpecName: "kube-api-access-9vhw5") pod "9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" (UID: "9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0"). InnerVolumeSpecName "kube-api-access-9vhw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.822831 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9vhw5\" (UniqueName: \"kubernetes.io/projected/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-kube-api-access-9vhw5\") on node \"crc\" DevicePath \"\"" Dec 03 17:26:35 crc kubenswrapper[4787]: I1203 17:26:35.823216 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2c8c\" (UniqueName: \"kubernetes.io/projected/74207f9d-ab04-4834-9813-5d89fe068d43-kube-api-access-n2c8c\") on node \"crc\" DevicePath \"\"" Dec 03 17:26:36 crc kubenswrapper[4787]: I1203 17:26:36.263330 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" event={"ID":"9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0","Type":"ContainerDied","Data":"51a0d6b980ddc846819ccf04a73ee18805b1a0be42bca177c9f6c7d88ecef6f2"} Dec 03 17:26:36 crc kubenswrapper[4787]: I1203 17:26:36.263449 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="51a0d6b980ddc846819ccf04a73ee18805b1a0be42bca177c9f6c7d88ecef6f2" Dec 03 17:26:36 crc kubenswrapper[4787]: I1203 17:26:36.263512 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j" Dec 03 17:26:36 crc kubenswrapper[4787]: I1203 17:26:36.267231 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" Dec 03 17:26:36 crc kubenswrapper[4787]: I1203 17:26:36.267400 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j" event={"ID":"74207f9d-ab04-4834-9813-5d89fe068d43","Type":"ContainerDied","Data":"0c4322a63d9089d8c2e4c6c8860275371cbd09a11bfa72a23b2a08f3aae8d3de"} Dec 03 17:26:36 crc kubenswrapper[4787]: I1203 17:26:36.267598 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c4322a63d9089d8c2e4c6c8860275371cbd09a11bfa72a23b2a08f3aae8d3de" Dec 03 17:26:36 crc kubenswrapper[4787]: I1203 17:26:36.531216 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-util" (OuterVolumeSpecName: "util") pod "74207f9d-ab04-4834-9813-5d89fe068d43" (UID: "74207f9d-ab04-4834-9813-5d89fe068d43"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:26:36 crc kubenswrapper[4787]: I1203 17:26:36.533907 4787 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/74207f9d-ab04-4834-9813-5d89fe068d43-util\") on node \"crc\" DevicePath \"\"" Dec 03 17:26:36 crc kubenswrapper[4787]: I1203 17:26:36.538489 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-util" (OuterVolumeSpecName: "util") pod "9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" (UID: "9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:26:36 crc kubenswrapper[4787]: I1203 17:26:36.636062 4787 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0-util\") on node \"crc\" DevicePath \"\"" Dec 03 17:26:37 crc kubenswrapper[4787]: I1203 17:26:37.276283 4787 generic.go:334] "Generic (PLEG): container finished" podID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerID="5fb5197a75036fd07fe1267a8e3f9b3739a483f7c55b1bd48a635bf808c22715" exitCode=0 Dec 03 17:26:37 crc kubenswrapper[4787]: I1203 17:26:37.276343 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k9kw4" event={"ID":"d075806c-3c60-4139-8ecf-9c9a384d1c24","Type":"ContainerDied","Data":"5fb5197a75036fd07fe1267a8e3f9b3739a483f7c55b1bd48a635bf808c22715"} Dec 03 17:26:38 crc kubenswrapper[4787]: I1203 17:26:38.287260 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k9kw4" event={"ID":"d075806c-3c60-4139-8ecf-9c9a384d1c24","Type":"ContainerStarted","Data":"de04a0d0c40be6be9ab7e78e6e540f6dab6474b37f1098e818928e4ccfc2ed78"} Dec 03 17:26:38 crc kubenswrapper[4787]: I1203 17:26:38.308902 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k9kw4" podStartSLOduration=2.827400366 podStartE2EDuration="6.308876202s" podCreationTimestamp="2025-12-03 17:26:32 +0000 UTC" firstStartedPulling="2025-12-03 17:26:34.249873054 +0000 UTC m=+811.067344013" lastFinishedPulling="2025-12-03 17:26:37.73134889 +0000 UTC m=+814.548819849" observedRunningTime="2025-12-03 17:26:38.305879819 +0000 UTC m=+815.123350788" watchObservedRunningTime="2025-12-03 17:26:38.308876202 +0000 UTC m=+815.126347171" Dec 03 17:26:43 crc kubenswrapper[4787]: I1203 17:26:43.092598 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:43 crc kubenswrapper[4787]: I1203 17:26:43.093136 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:44 crc kubenswrapper[4787]: I1203 17:26:44.131446 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k9kw4" podUID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerName="registry-server" probeResult="failure" output=< Dec 03 17:26:44 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 17:26:44 crc kubenswrapper[4787]: > Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.093982 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll"] Dec 03 17:26:46 crc kubenswrapper[4787]: E1203 17:26:46.094737 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74207f9d-ab04-4834-9813-5d89fe068d43" containerName="util" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.094756 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="74207f9d-ab04-4834-9813-5d89fe068d43" containerName="util" Dec 03 17:26:46 crc kubenswrapper[4787]: E1203 17:26:46.094765 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74207f9d-ab04-4834-9813-5d89fe068d43" containerName="pull" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.094773 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="74207f9d-ab04-4834-9813-5d89fe068d43" containerName="pull" Dec 03 17:26:46 crc kubenswrapper[4787]: E1203 17:26:46.094787 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" containerName="util" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.094795 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" containerName="util" Dec 03 17:26:46 crc kubenswrapper[4787]: E1203 17:26:46.094811 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" containerName="pull" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.094818 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" containerName="pull" Dec 03 17:26:46 crc kubenswrapper[4787]: E1203 17:26:46.094826 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" containerName="extract" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.094834 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" containerName="extract" Dec 03 17:26:46 crc kubenswrapper[4787]: E1203 17:26:46.094849 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74207f9d-ab04-4834-9813-5d89fe068d43" containerName="extract" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.094857 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="74207f9d-ab04-4834-9813-5d89fe068d43" containerName="extract" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.094984 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="74207f9d-ab04-4834-9813-5d89fe068d43" containerName="extract" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.095002 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0" containerName="extract" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.095774 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.098831 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-dockercfg-47dz7" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.099324 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-controller-manager-service-cert" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.099599 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators-redhat"/"loki-operator-metrics" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.099821 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"openshift-service-ca.crt" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.100165 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"kube-root-ca.crt" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.100387 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators-redhat"/"loki-operator-manager-config" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.110443 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll"] Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.279934 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c73f5b31-0cde-406e-800b-153981eeead6-webhook-cert\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.279995 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/c73f5b31-0cde-406e-800b-153981eeead6-manager-config\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.280110 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c73f5b31-0cde-406e-800b-153981eeead6-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.280150 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lc7xf\" (UniqueName: \"kubernetes.io/projected/c73f5b31-0cde-406e-800b-153981eeead6-kube-api-access-lc7xf\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.280196 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c73f5b31-0cde-406e-800b-153981eeead6-apiservice-cert\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.381459 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/c73f5b31-0cde-406e-800b-153981eeead6-manager-config\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.381544 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c73f5b31-0cde-406e-800b-153981eeead6-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.381573 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lc7xf\" (UniqueName: \"kubernetes.io/projected/c73f5b31-0cde-406e-800b-153981eeead6-kube-api-access-lc7xf\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.381615 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c73f5b31-0cde-406e-800b-153981eeead6-apiservice-cert\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.381665 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c73f5b31-0cde-406e-800b-153981eeead6-webhook-cert\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.383692 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manager-config\" (UniqueName: \"kubernetes.io/configmap/c73f5b31-0cde-406e-800b-153981eeead6-manager-config\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.390296 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c73f5b31-0cde-406e-800b-153981eeead6-apiservice-cert\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.390965 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c73f5b31-0cde-406e-800b-153981eeead6-webhook-cert\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.404287 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"loki-operator-metrics-cert\" (UniqueName: \"kubernetes.io/secret/c73f5b31-0cde-406e-800b-153981eeead6-loki-operator-metrics-cert\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.407089 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lc7xf\" (UniqueName: \"kubernetes.io/projected/c73f5b31-0cde-406e-800b-153981eeead6-kube-api-access-lc7xf\") pod \"loki-operator-controller-manager-5cd867f87d-zwxll\" (UID: \"c73f5b31-0cde-406e-800b-153981eeead6\") " pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.414803 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:26:46 crc kubenswrapper[4787]: I1203 17:26:46.883149 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll"] Dec 03 17:26:47 crc kubenswrapper[4787]: I1203 17:26:47.339642 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" event={"ID":"c73f5b31-0cde-406e-800b-153981eeead6","Type":"ContainerStarted","Data":"22e9fe4c23b6516f224e94c604dcc4dfe528bf109da4372ea3602ea799d74157"} Dec 03 17:26:48 crc kubenswrapper[4787]: I1203 17:26:48.990603 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:26:48 crc kubenswrapper[4787]: I1203 17:26:48.990982 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:26:49 crc kubenswrapper[4787]: I1203 17:26:49.995246 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-z7f2j"] Dec 03 17:26:49 crc kubenswrapper[4787]: I1203 17:26:49.996082 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-z7f2j" Dec 03 17:26:49 crc kubenswrapper[4787]: I1203 17:26:49.999092 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"cluster-logging-operator-dockercfg-bv7kw" Dec 03 17:26:49 crc kubenswrapper[4787]: I1203 17:26:49.999093 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"openshift-service-ca.crt" Dec 03 17:26:50 crc kubenswrapper[4787]: I1203 17:26:50.001799 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"kube-root-ca.crt" Dec 03 17:26:50 crc kubenswrapper[4787]: I1203 17:26:50.015733 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-z7f2j"] Dec 03 17:26:50 crc kubenswrapper[4787]: I1203 17:26:50.143342 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k2j4j\" (UniqueName: \"kubernetes.io/projected/28f7de1d-ea99-4b90-bdb9-bc1f04f201ff-kube-api-access-k2j4j\") pod \"cluster-logging-operator-ff9846bd-z7f2j\" (UID: \"28f7de1d-ea99-4b90-bdb9-bc1f04f201ff\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-z7f2j" Dec 03 17:26:50 crc kubenswrapper[4787]: I1203 17:26:50.244753 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k2j4j\" (UniqueName: \"kubernetes.io/projected/28f7de1d-ea99-4b90-bdb9-bc1f04f201ff-kube-api-access-k2j4j\") pod \"cluster-logging-operator-ff9846bd-z7f2j\" (UID: \"28f7de1d-ea99-4b90-bdb9-bc1f04f201ff\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-z7f2j" Dec 03 17:26:50 crc kubenswrapper[4787]: I1203 17:26:50.315409 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k2j4j\" (UniqueName: \"kubernetes.io/projected/28f7de1d-ea99-4b90-bdb9-bc1f04f201ff-kube-api-access-k2j4j\") pod \"cluster-logging-operator-ff9846bd-z7f2j\" (UID: \"28f7de1d-ea99-4b90-bdb9-bc1f04f201ff\") " pod="openshift-logging/cluster-logging-operator-ff9846bd-z7f2j" Dec 03 17:26:50 crc kubenswrapper[4787]: I1203 17:26:50.380891 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/cluster-logging-operator-ff9846bd-z7f2j" Dec 03 17:26:53 crc kubenswrapper[4787]: I1203 17:26:53.152082 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:53 crc kubenswrapper[4787]: I1203 17:26:53.222636 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:26:53 crc kubenswrapper[4787]: I1203 17:26:53.831264 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/cluster-logging-operator-ff9846bd-z7f2j"] Dec 03 17:26:54 crc kubenswrapper[4787]: I1203 17:26:54.385576 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-z7f2j" event={"ID":"28f7de1d-ea99-4b90-bdb9-bc1f04f201ff","Type":"ContainerStarted","Data":"1a637c3140ab3ced4b2ccbd975f7a6fdb3ff4ef3a21dd08be7a6c5d94fc785f0"} Dec 03 17:26:54 crc kubenswrapper[4787]: I1203 17:26:54.386927 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" event={"ID":"c73f5b31-0cde-406e-800b-153981eeead6","Type":"ContainerStarted","Data":"c9ad861f1e54304fee67486a37eb11dcccac2e943ebe6d9709be2c1727f79d93"} Dec 03 17:26:56 crc kubenswrapper[4787]: I1203 17:26:56.290681 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k9kw4"] Dec 03 17:26:56 crc kubenswrapper[4787]: I1203 17:26:56.291180 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k9kw4" podUID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerName="registry-server" containerID="cri-o://de04a0d0c40be6be9ab7e78e6e540f6dab6474b37f1098e818928e4ccfc2ed78" gracePeriod=2 Dec 03 17:26:57 crc kubenswrapper[4787]: I1203 17:26:57.500637 4787 generic.go:334] "Generic (PLEG): container finished" podID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerID="de04a0d0c40be6be9ab7e78e6e540f6dab6474b37f1098e818928e4ccfc2ed78" exitCode=0 Dec 03 17:26:57 crc kubenswrapper[4787]: I1203 17:26:57.502669 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k9kw4" event={"ID":"d075806c-3c60-4139-8ecf-9c9a384d1c24","Type":"ContainerDied","Data":"de04a0d0c40be6be9ab7e78e6e540f6dab6474b37f1098e818928e4ccfc2ed78"} Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.459036 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.526376 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k9kw4" event={"ID":"d075806c-3c60-4139-8ecf-9c9a384d1c24","Type":"ContainerDied","Data":"3ab1892ec91374654437d1859136b601a7e5d5f9d0d50f006bc73411bf8d951c"} Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.526452 4787 scope.go:117] "RemoveContainer" containerID="de04a0d0c40be6be9ab7e78e6e540f6dab6474b37f1098e818928e4ccfc2ed78" Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.526492 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k9kw4" Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.585826 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-catalog-content\") pod \"d075806c-3c60-4139-8ecf-9c9a384d1c24\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.585943 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6c2p\" (UniqueName: \"kubernetes.io/projected/d075806c-3c60-4139-8ecf-9c9a384d1c24-kube-api-access-b6c2p\") pod \"d075806c-3c60-4139-8ecf-9c9a384d1c24\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.586050 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-utilities\") pod \"d075806c-3c60-4139-8ecf-9c9a384d1c24\" (UID: \"d075806c-3c60-4139-8ecf-9c9a384d1c24\") " Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.587433 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-utilities" (OuterVolumeSpecName: "utilities") pod "d075806c-3c60-4139-8ecf-9c9a384d1c24" (UID: "d075806c-3c60-4139-8ecf-9c9a384d1c24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.593297 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d075806c-3c60-4139-8ecf-9c9a384d1c24-kube-api-access-b6c2p" (OuterVolumeSpecName: "kube-api-access-b6c2p") pod "d075806c-3c60-4139-8ecf-9c9a384d1c24" (UID: "d075806c-3c60-4139-8ecf-9c9a384d1c24"). InnerVolumeSpecName "kube-api-access-b6c2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.688382 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6c2p\" (UniqueName: \"kubernetes.io/projected/d075806c-3c60-4139-8ecf-9c9a384d1c24-kube-api-access-b6c2p\") on node \"crc\" DevicePath \"\"" Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.688428 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.720617 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d075806c-3c60-4139-8ecf-9c9a384d1c24" (UID: "d075806c-3c60-4139-8ecf-9c9a384d1c24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.789934 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d075806c-3c60-4139-8ecf-9c9a384d1c24-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.856140 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k9kw4"] Dec 03 17:27:00 crc kubenswrapper[4787]: I1203 17:27:00.859918 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k9kw4"] Dec 03 17:27:01 crc kubenswrapper[4787]: I1203 17:27:01.773643 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d075806c-3c60-4139-8ecf-9c9a384d1c24" path="/var/lib/kubelet/pods/d075806c-3c60-4139-8ecf-9c9a384d1c24/volumes" Dec 03 17:27:02 crc kubenswrapper[4787]: I1203 17:27:02.387461 4787 scope.go:117] "RemoveContainer" containerID="5fb5197a75036fd07fe1267a8e3f9b3739a483f7c55b1bd48a635bf808c22715" Dec 03 17:27:02 crc kubenswrapper[4787]: I1203 17:27:02.509148 4787 scope.go:117] "RemoveContainer" containerID="de9ec3f2fc3b8cb05db0812ee10c5f1f9559ee31403d271ae968254b655ac3bb" Dec 03 17:27:03 crc kubenswrapper[4787]: I1203 17:27:03.550274 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/cluster-logging-operator-ff9846bd-z7f2j" event={"ID":"28f7de1d-ea99-4b90-bdb9-bc1f04f201ff","Type":"ContainerStarted","Data":"e76316f785838112dc4f37e21f2c29582ce47b610435b7c70d641abea5b4f4e3"} Dec 03 17:27:03 crc kubenswrapper[4787]: I1203 17:27:03.552330 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" event={"ID":"c73f5b31-0cde-406e-800b-153981eeead6","Type":"ContainerStarted","Data":"3eae811e2c4bdc7794ae0d0c8bef87b3f0f5dc6291f156680372cccad5039613"} Dec 03 17:27:03 crc kubenswrapper[4787]: I1203 17:27:03.552792 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:27:03 crc kubenswrapper[4787]: I1203 17:27:03.554715 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" Dec 03 17:27:03 crc kubenswrapper[4787]: I1203 17:27:03.569282 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/cluster-logging-operator-ff9846bd-z7f2j" podStartSLOduration=5.87900691 podStartE2EDuration="14.56926036s" podCreationTimestamp="2025-12-03 17:26:49 +0000 UTC" firstStartedPulling="2025-12-03 17:26:53.840058647 +0000 UTC m=+830.657529606" lastFinishedPulling="2025-12-03 17:27:02.530312097 +0000 UTC m=+839.347783056" observedRunningTime="2025-12-03 17:27:03.565107234 +0000 UTC m=+840.382578203" watchObservedRunningTime="2025-12-03 17:27:03.56926036 +0000 UTC m=+840.386731319" Dec 03 17:27:03 crc kubenswrapper[4787]: I1203 17:27:03.589037 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators-redhat/loki-operator-controller-manager-5cd867f87d-zwxll" podStartSLOduration=1.9526773309999998 podStartE2EDuration="17.58900033s" podCreationTimestamp="2025-12-03 17:26:46 +0000 UTC" firstStartedPulling="2025-12-03 17:26:46.892519917 +0000 UTC m=+823.709990876" lastFinishedPulling="2025-12-03 17:27:02.528842916 +0000 UTC m=+839.346313875" observedRunningTime="2025-12-03 17:27:03.587966829 +0000 UTC m=+840.405437818" watchObservedRunningTime="2025-12-03 17:27:03.58900033 +0000 UTC m=+840.406471289" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.702785 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["minio-dev/minio"] Dec 03 17:27:08 crc kubenswrapper[4787]: E1203 17:27:08.704405 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerName="extract-utilities" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.704499 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerName="extract-utilities" Dec 03 17:27:08 crc kubenswrapper[4787]: E1203 17:27:08.704573 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerName="registry-server" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.704642 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerName="registry-server" Dec 03 17:27:08 crc kubenswrapper[4787]: E1203 17:27:08.704722 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerName="extract-content" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.704796 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerName="extract-content" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.704977 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d075806c-3c60-4139-8ecf-9c9a384d1c24" containerName="registry-server" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.705576 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.708158 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"kube-root-ca.crt" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.751090 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"minio-dev"/"openshift-service-ca.crt" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.753382 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.796002 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-00c76780-bd85-483a-9417-288294d73527\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00c76780-bd85-483a-9417-288294d73527\") pod \"minio\" (UID: \"106cf5d0-b4eb-4fdb-8769-d893ce141427\") " pod="minio-dev/minio" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.796133 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nzvrk\" (UniqueName: \"kubernetes.io/projected/106cf5d0-b4eb-4fdb-8769-d893ce141427-kube-api-access-nzvrk\") pod \"minio\" (UID: \"106cf5d0-b4eb-4fdb-8769-d893ce141427\") " pod="minio-dev/minio" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.897408 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-00c76780-bd85-483a-9417-288294d73527\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00c76780-bd85-483a-9417-288294d73527\") pod \"minio\" (UID: \"106cf5d0-b4eb-4fdb-8769-d893ce141427\") " pod="minio-dev/minio" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.897496 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nzvrk\" (UniqueName: \"kubernetes.io/projected/106cf5d0-b4eb-4fdb-8769-d893ce141427-kube-api-access-nzvrk\") pod \"minio\" (UID: \"106cf5d0-b4eb-4fdb-8769-d893ce141427\") " pod="minio-dev/minio" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.902418 4787 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.902471 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-00c76780-bd85-483a-9417-288294d73527\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00c76780-bd85-483a-9417-288294d73527\") pod \"minio\" (UID: \"106cf5d0-b4eb-4fdb-8769-d893ce141427\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/498af5cc2d58200fb4e577f492a5db14d50832c497f80d3fcce7f799eb3d9937/globalmount\"" pod="minio-dev/minio" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.921709 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nzvrk\" (UniqueName: \"kubernetes.io/projected/106cf5d0-b4eb-4fdb-8769-d893ce141427-kube-api-access-nzvrk\") pod \"minio\" (UID: \"106cf5d0-b4eb-4fdb-8769-d893ce141427\") " pod="minio-dev/minio" Dec 03 17:27:08 crc kubenswrapper[4787]: I1203 17:27:08.932003 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-00c76780-bd85-483a-9417-288294d73527\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-00c76780-bd85-483a-9417-288294d73527\") pod \"minio\" (UID: \"106cf5d0-b4eb-4fdb-8769-d893ce141427\") " pod="minio-dev/minio" Dec 03 17:27:09 crc kubenswrapper[4787]: I1203 17:27:09.065255 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="minio-dev/minio" Dec 03 17:27:09 crc kubenswrapper[4787]: I1203 17:27:09.499408 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["minio-dev/minio"] Dec 03 17:27:09 crc kubenswrapper[4787]: I1203 17:27:09.584756 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"106cf5d0-b4eb-4fdb-8769-d893ce141427","Type":"ContainerStarted","Data":"710359d47a5c6bdd8db6669c6495da71b5107bdec2479ffa70b9788a4e7b7c30"} Dec 03 17:27:13 crc kubenswrapper[4787]: I1203 17:27:13.629891 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="minio-dev/minio" event={"ID":"106cf5d0-b4eb-4fdb-8769-d893ce141427","Type":"ContainerStarted","Data":"ee555e3d68f7153431774f204dfdaba0d72f288e8dbf7d2ea4b70903db600ae2"} Dec 03 17:27:13 crc kubenswrapper[4787]: I1203 17:27:13.652331 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="minio-dev/minio" podStartSLOduration=4.5379431740000005 podStartE2EDuration="7.652299696s" podCreationTimestamp="2025-12-03 17:27:06 +0000 UTC" firstStartedPulling="2025-12-03 17:27:09.513563056 +0000 UTC m=+846.331034015" lastFinishedPulling="2025-12-03 17:27:12.627919568 +0000 UTC m=+849.445390537" observedRunningTime="2025-12-03 17:27:13.643984167 +0000 UTC m=+850.461455126" watchObservedRunningTime="2025-12-03 17:27:13.652299696 +0000 UTC m=+850.469770655" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.068989 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-flghj"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.071338 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.076755 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-ca-bundle" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.077183 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-config" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.077474 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-grpc" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.077485 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-dockercfg-s46vx" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.077837 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-distributor-http" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.099074 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-flghj"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.226527 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-xq92s"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.227993 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.231317 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-grpc" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.231894 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-querier-http" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.232772 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-s3" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.241417 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/e1be9303-41d5-431a-bb49-b0e104ce4625-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.241551 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e1be9303-41d5-431a-bb49-b0e104ce4625-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.241611 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1be9303-41d5-431a-bb49-b0e104ce4625-config\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.241630 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1be9303-41d5-431a-bb49-b0e104ce4625-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.241795 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhb7p\" (UniqueName: \"kubernetes.io/projected/e1be9303-41d5-431a-bb49-b0e104ce4625-kube-api-access-dhb7p\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.248747 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-xq92s"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.328865 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.331045 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.334107 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-http" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.334587 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-query-frontend-grpc" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.342595 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2bg8\" (UniqueName: \"kubernetes.io/projected/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-kube-api-access-w2bg8\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.342652 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhb7p\" (UniqueName: \"kubernetes.io/projected/e1be9303-41d5-431a-bb49-b0e104ce4625-kube-api-access-dhb7p\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.342683 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.342722 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.342745 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/e1be9303-41d5-431a-bb49-b0e104ce4625-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.342765 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-config\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.342786 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.342826 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e1be9303-41d5-431a-bb49-b0e104ce4625-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.342860 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1be9303-41d5-431a-bb49-b0e104ce4625-config\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.342880 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1be9303-41d5-431a-bb49-b0e104ce4625-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.342902 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.344427 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e1be9303-41d5-431a-bb49-b0e104ce4625-config\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.345166 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e1be9303-41d5-431a-bb49-b0e104ce4625-logging-loki-ca-bundle\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.345471 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.353242 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-grpc\" (UniqueName: \"kubernetes.io/secret/e1be9303-41d5-431a-bb49-b0e104ce4625-logging-loki-distributor-grpc\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.361355 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-distributor-http\" (UniqueName: \"kubernetes.io/secret/e1be9303-41d5-431a-bb49-b0e104ce4625-logging-loki-distributor-http\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.369060 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhb7p\" (UniqueName: \"kubernetes.io/projected/e1be9303-41d5-431a-bb49-b0e104ce4625-kube-api-access-dhb7p\") pod \"logging-loki-distributor-76cc67bf56-flghj\" (UID: \"e1be9303-41d5-431a-bb49-b0e104ce4625\") " pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.402508 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444225 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444446 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2bg8\" (UniqueName: \"kubernetes.io/projected/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-kube-api-access-w2bg8\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444504 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444536 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-config\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444604 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444636 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444678 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-config\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444736 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444817 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444857 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444919 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.444956 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh7l6\" (UniqueName: \"kubernetes.io/projected/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-kube-api-access-dh7l6\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.446056 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-config\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.446440 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.448116 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-ca-bundle\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.454590 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-client-http" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.454834 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.454964 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway-ca-bundle" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.455108 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-s3\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.455237 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-http" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.455687 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"logging-loki-gateway" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.455904 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-http\" (UniqueName: \"kubernetes.io/secret/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-querier-http\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.456517 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-querier-grpc\" (UniqueName: \"kubernetes.io/secret/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-logging-loki-querier-grpc\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.460803 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.462281 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.464577 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-gateway-dockercfg-qfwc9" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.473288 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2bg8\" (UniqueName: \"kubernetes.io/projected/4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a-kube-api-access-w2bg8\") pod \"logging-loki-querier-5895d59bb8-xq92s\" (UID: \"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a\") " pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.485931 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.490861 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.541818 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546514 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh7l6\" (UniqueName: \"kubernetes.io/projected/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-kube-api-access-dh7l6\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546558 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-tenants\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546584 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-lokistack-gateway\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546617 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-config\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546639 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-rbac\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546662 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546683 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546702 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-tls-secret\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546718 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lsrf4\" (UniqueName: \"kubernetes.io/projected/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-kube-api-access-lsrf4\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546737 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546760 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546791 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.546817 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-logging-loki-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.548657 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-logging-loki-ca-bundle\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.549388 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-config\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.555525 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-http\" (UniqueName: \"kubernetes.io/secret/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-logging-loki-query-frontend-http\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.555679 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-query-frontend-grpc\" (UniqueName: \"kubernetes.io/secret/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-logging-loki-query-frontend-grpc\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.570315 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh7l6\" (UniqueName: \"kubernetes.io/projected/b43db9b8-11f8-4fb7-82d2-6b38e34e1a02-kube-api-access-dh7l6\") pod \"logging-loki-query-frontend-84558f7c9f-zsjlj\" (UID: \"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02\") " pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.647175 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648506 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-logging-loki-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648595 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-tenants\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648638 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk8vh\" (UniqueName: \"kubernetes.io/projected/cb532cad-464c-4fe6-99de-b7897de9ac51-kube-api-access-qk8vh\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648669 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648694 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-lokistack-gateway\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648731 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-tenants\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648771 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-tls-secret\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648803 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648826 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-logging-loki-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648855 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-rbac\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648878 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-lokistack-gateway\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648901 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648922 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-tls-secret\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648939 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lsrf4\" (UniqueName: \"kubernetes.io/projected/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-kube-api-access-lsrf4\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648960 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.648983 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-rbac\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.649693 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-logging-loki-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.649808 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-lokistack-gateway\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: E1203 17:27:18.649813 4787 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Dec 03 17:27:18 crc kubenswrapper[4787]: E1203 17:27:18.649891 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-tls-secret podName:35486d7a-dcc2-40bf-b2b0-869ecfe46ccc nodeName:}" failed. No retries permitted until 2025-12-03 17:27:19.14987489 +0000 UTC m=+855.967345849 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-tls-secret") pod "logging-loki-gateway-57f76f4c5d-dzw77" (UID: "35486d7a-dcc2-40bf-b2b0-869ecfe46ccc") : secret "logging-loki-gateway-http" not found Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.653052 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.653063 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-tenants\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.653593 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-rbac\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.655807 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.675707 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lsrf4\" (UniqueName: \"kubernetes.io/projected/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-kube-api-access-lsrf4\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.753726 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-tls-secret\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.753776 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.753795 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-logging-loki-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.753819 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-lokistack-gateway\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.753870 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-rbac\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: E1203 17:27:18.753896 4787 secret.go:188] Couldn't get secret openshift-logging/logging-loki-gateway-http: secret "logging-loki-gateway-http" not found Dec 03 17:27:18 crc kubenswrapper[4787]: E1203 17:27:18.753967 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-tls-secret podName:cb532cad-464c-4fe6-99de-b7897de9ac51 nodeName:}" failed. No retries permitted until 2025-12-03 17:27:19.25394964 +0000 UTC m=+856.071420589 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-secret" (UniqueName: "kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-tls-secret") pod "logging-loki-gateway-57f76f4c5d-v45xd" (UID: "cb532cad-464c-4fe6-99de-b7897de9ac51") : secret "logging-loki-gateway-http" not found Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.753908 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk8vh\" (UniqueName: \"kubernetes.io/projected/cb532cad-464c-4fe6-99de-b7897de9ac51-kube-api-access-qk8vh\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.754220 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.754733 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-logging-loki-gateway-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.755448 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lokistack-gateway\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-lokistack-gateway\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.761600 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-gateway-client-http\" (UniqueName: \"kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-logging-loki-gateway-client-http\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.762285 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-logging-loki-ca-bundle\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.762391 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-tenants\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.764246 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rbac\" (UniqueName: \"kubernetes.io/configmap/cb532cad-464c-4fe6-99de-b7897de9ac51-rbac\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.768832 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tenants\" (UniqueName: \"kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-tenants\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.771986 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk8vh\" (UniqueName: \"kubernetes.io/projected/cb532cad-464c-4fe6-99de-b7897de9ac51-kube-api-access-qk8vh\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.845116 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-querier-5895d59bb8-xq92s"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.943955 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-distributor-76cc67bf56-flghj"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.981158 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj"] Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.989987 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.990078 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.990139 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.991040 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"49596b58389f8b6bbf73d259649cd03b2e93ae7e21af7479d7bc4161137e279b"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:27:18 crc kubenswrapper[4787]: I1203 17:27:18.991113 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://49596b58389f8b6bbf73d259649cd03b2e93ae7e21af7479d7bc4161137e279b" gracePeriod=600 Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.173331 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-tls-secret\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.185271 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/35486d7a-dcc2-40bf-b2b0-869ecfe46ccc-tls-secret\") pod \"logging-loki-gateway-57f76f4c5d-dzw77\" (UID: \"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.275422 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-tls-secret\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.278652 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.282088 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.283684 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-secret\" (UniqueName: \"kubernetes.io/secret/cb532cad-464c-4fe6-99de-b7897de9ac51-tls-secret\") pod \"logging-loki-gateway-57f76f4c5d-v45xd\" (UID: \"cb532cad-464c-4fe6-99de-b7897de9ac51\") " pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.286616 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-http" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.286845 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-ingester-grpc" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.298321 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.303961 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.304740 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.307705 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-grpc" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.307838 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-compactor-http" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.348509 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.377590 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.377646 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-0f5239a9-7c63-441c-aa8d-9ffc7519eee6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0f5239a9-7c63-441c-aa8d-9ffc7519eee6\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.377686 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.377705 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.377731 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-fae78d4b-fa8f-4900-b212-5262f0d6e4b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fae78d4b-fa8f-4900-b212-5262f0d6e4b0\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.377750 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mslvg\" (UniqueName: \"kubernetes.io/projected/7da69ee7-38ad-4cce-9bf0-140e7283a775-kube-api-access-mslvg\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.377783 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.377809 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da69ee7-38ad-4cce-9bf0-140e7283a775-config\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.390445 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.406260 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.406526 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.415361 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-grpc" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.415556 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"logging-loki-index-gateway-http" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.419466 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.423062 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.485505 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486082 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-0f5239a9-7c63-441c-aa8d-9ffc7519eee6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0f5239a9-7c63-441c-aa8d-9ffc7519eee6\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486118 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486158 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-3e371fd7-a412-4695-96c9-ea91ac71a3db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3e371fd7-a412-4695-96c9-ea91ac71a3db\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486182 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486204 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486227 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486250 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hb7f5\" (UniqueName: \"kubernetes.io/projected/db4243ab-4a52-4991-89a6-96e1ef9a348f-kube-api-access-hb7f5\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486305 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486354 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-fae78d4b-fa8f-4900-b212-5262f0d6e4b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fae78d4b-fa8f-4900-b212-5262f0d6e4b0\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486386 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486418 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mslvg\" (UniqueName: \"kubernetes.io/projected/7da69ee7-38ad-4cce-9bf0-140e7283a775-kube-api-access-mslvg\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486454 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db4243ab-4a52-4991-89a6-96e1ef9a348f-config\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486483 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-ddf1c201-ac93-4e96-91da-b837a57e41ba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ddf1c201-ac93-4e96-91da-b837a57e41ba\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486522 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5j4w2\" (UniqueName: \"kubernetes.io/projected/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-kube-api-access-5j4w2\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486549 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486576 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486599 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-config\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486629 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da69ee7-38ad-4cce-9bf0-140e7283a775-config\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486652 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486689 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.486715 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.499700 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7da69ee7-38ad-4cce-9bf0-140e7283a775-config\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.500261 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-ca-bundle\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.506305 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-grpc\" (UniqueName: \"kubernetes.io/secret/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-ingester-grpc\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.507106 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-s3\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.515890 4787 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.516075 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-0f5239a9-7c63-441c-aa8d-9ffc7519eee6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0f5239a9-7c63-441c-aa8d-9ffc7519eee6\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/ff209e5a5555538227f74fb85da3d7e8239d2fb32b21567fa0a31f2237438472/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.518326 4787 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.518406 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-fae78d4b-fa8f-4900-b212-5262f0d6e4b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fae78d4b-fa8f-4900-b212-5262f0d6e4b0\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f483903201c338a774300e675dd2a11db29a03efd27e3e4c216b887cbb9dbdb0/globalmount\"" pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.520840 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ingester-http\" (UniqueName: \"kubernetes.io/secret/7da69ee7-38ad-4cce-9bf0-140e7283a775-logging-loki-ingester-http\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.530987 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mslvg\" (UniqueName: \"kubernetes.io/projected/7da69ee7-38ad-4cce-9bf0-140e7283a775-kube-api-access-mslvg\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.588860 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.589236 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.589340 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.589475 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.589575 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-3e371fd7-a412-4695-96c9-ea91ac71a3db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3e371fd7-a412-4695-96c9-ea91ac71a3db\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.589736 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.589846 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hb7f5\" (UniqueName: \"kubernetes.io/projected/db4243ab-4a52-4991-89a6-96e1ef9a348f-kube-api-access-hb7f5\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.589944 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.590091 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.590204 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db4243ab-4a52-4991-89a6-96e1ef9a348f-config\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.601119 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-ddf1c201-ac93-4e96-91da-b837a57e41ba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ddf1c201-ac93-4e96-91da-b837a57e41ba\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.601256 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5j4w2\" (UniqueName: \"kubernetes.io/projected/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-kube-api-access-5j4w2\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.601310 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.601339 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-config\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.594533 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/db4243ab-4a52-4991-89a6-96e1ef9a348f-config\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.590137 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-ca-bundle\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.595468 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-grpc\" (UniqueName: \"kubernetes.io/secret/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-index-gateway-grpc\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.590544 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-0f5239a9-7c63-441c-aa8d-9ffc7519eee6\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-0f5239a9-7c63-441c-aa8d-9ffc7519eee6\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.592780 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-ca-bundle\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.593286 4787 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.602974 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-3e371fd7-a412-4695-96c9-ea91ac71a3db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3e371fd7-a412-4695-96c9-ea91ac71a3db\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/f0e549b79e4e84b7aab40305dad1d8d65091e2a12a8bbed6f10159e7840ce251/globalmount\"" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.598938 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-fae78d4b-fa8f-4900-b212-5262f0d6e4b0\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-fae78d4b-fa8f-4900-b212-5262f0d6e4b0\") pod \"logging-loki-ingester-0\" (UID: \"7da69ee7-38ad-4cce-9bf0-140e7283a775\") " pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.606624 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-grpc\" (UniqueName: \"kubernetes.io/secret/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-compactor-grpc\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.606672 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-index-gateway-http\" (UniqueName: \"kubernetes.io/secret/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-index-gateway-http\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.594556 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-s3\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.603945 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-config\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.611904 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-compactor-http\" (UniqueName: \"kubernetes.io/secret/db4243ab-4a52-4991-89a6-96e1ef9a348f-logging-loki-compactor-http\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.612130 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hb7f5\" (UniqueName: \"kubernetes.io/projected/db4243ab-4a52-4991-89a6-96e1ef9a348f-kube-api-access-hb7f5\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.616570 4787 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.616646 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-ddf1c201-ac93-4e96-91da-b837a57e41ba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ddf1c201-ac93-4e96-91da-b837a57e41ba\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/09788b5fe67e21d0c4d92caebc0e6bf2defe9965ded42d13254a15ef804702df/globalmount\"" pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.620293 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-loki-s3\" (UniqueName: \"kubernetes.io/secret/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-logging-loki-s3\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.653173 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5j4w2\" (UniqueName: \"kubernetes.io/projected/b4de518a-6a4d-441b-b04c-d35ca9a8b5f9-kube-api-access-5j4w2\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.664801 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-3e371fd7-a412-4695-96c9-ea91ac71a3db\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-3e371fd7-a412-4695-96c9-ea91ac71a3db\") pod \"logging-loki-index-gateway-0\" (UID: \"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9\") " pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.669704 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-ddf1c201-ac93-4e96-91da-b837a57e41ba\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-ddf1c201-ac93-4e96-91da-b837a57e41ba\") pod \"logging-loki-compactor-0\" (UID: \"db4243ab-4a52-4991-89a6-96e1ef9a348f\") " pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.684456 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" event={"ID":"e1be9303-41d5-431a-bb49-b0e104ce4625","Type":"ContainerStarted","Data":"451e18ac0c635e13e09da5488c13cfd458324a165008b3fc51ce096049c0b0e2"} Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.687115 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="49596b58389f8b6bbf73d259649cd03b2e93ae7e21af7479d7bc4161137e279b" exitCode=0 Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.687193 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"49596b58389f8b6bbf73d259649cd03b2e93ae7e21af7479d7bc4161137e279b"} Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.687331 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"864b789057addccde661186743b5963a0f8e318c2a7a639d2b92c72f83cc12f6"} Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.687364 4787 scope.go:117] "RemoveContainer" containerID="971d7c10a947bd7777cac22b6df04c66ed8ba415706fe05793b9542a5980006b" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.688354 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" event={"ID":"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02","Type":"ContainerStarted","Data":"4b6e19c44016b85298be6349ea62e4913988a90da0f4fc57c7d6eefa6481d40d"} Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.699629 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" event={"ID":"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a","Type":"ContainerStarted","Data":"975e4a8b5c58c1725c6145a4db228cb3ae8655e6504d6c025cfc8074236810d8"} Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.879660 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd"] Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.903545 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.907441 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.919270 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:19 crc kubenswrapper[4787]: I1203 17:27:19.984409 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77"] Dec 03 17:27:20 crc kubenswrapper[4787]: W1203 17:27:20.011136 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35486d7a_dcc2_40bf_b2b0_869ecfe46ccc.slice/crio-8d5e10b07b7770764f9152511cfdc7c1a4cfed7a1efbecca752b88035c9cd489 WatchSource:0}: Error finding container 8d5e10b07b7770764f9152511cfdc7c1a4cfed7a1efbecca752b88035c9cd489: Status 404 returned error can't find the container with id 8d5e10b07b7770764f9152511cfdc7c1a4cfed7a1efbecca752b88035c9cd489 Dec 03 17:27:20 crc kubenswrapper[4787]: I1203 17:27:20.484738 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-ingester-0"] Dec 03 17:27:20 crc kubenswrapper[4787]: W1203 17:27:20.515340 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7da69ee7_38ad_4cce_9bf0_140e7283a775.slice/crio-fa6844654ff5845c695c87864c6e6f0aadc1bea36b028a8927a246c797080a74 WatchSource:0}: Error finding container fa6844654ff5845c695c87864c6e6f0aadc1bea36b028a8927a246c797080a74: Status 404 returned error can't find the container with id fa6844654ff5845c695c87864c6e6f0aadc1bea36b028a8927a246c797080a74 Dec 03 17:27:20 crc kubenswrapper[4787]: I1203 17:27:20.578442 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-compactor-0"] Dec 03 17:27:20 crc kubenswrapper[4787]: I1203 17:27:20.588592 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/logging-loki-index-gateway-0"] Dec 03 17:27:20 crc kubenswrapper[4787]: W1203 17:27:20.624593 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb4de518a_6a4d_441b_b04c_d35ca9a8b5f9.slice/crio-9f241108687bb6e04e230409b1c9e4b4fddf6514f61b231103a7c8dd3186ab26 WatchSource:0}: Error finding container 9f241108687bb6e04e230409b1c9e4b4fddf6514f61b231103a7c8dd3186ab26: Status 404 returned error can't find the container with id 9f241108687bb6e04e230409b1c9e4b4fddf6514f61b231103a7c8dd3186ab26 Dec 03 17:27:20 crc kubenswrapper[4787]: I1203 17:27:20.709253 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"db4243ab-4a52-4991-89a6-96e1ef9a348f","Type":"ContainerStarted","Data":"dd9fb256e51eaae59c35e7f0cccc8cd18f85433793765cfb2bf0d5f7e2c8d173"} Dec 03 17:27:20 crc kubenswrapper[4787]: I1203 17:27:20.711614 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" event={"ID":"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc","Type":"ContainerStarted","Data":"8d5e10b07b7770764f9152511cfdc7c1a4cfed7a1efbecca752b88035c9cd489"} Dec 03 17:27:20 crc kubenswrapper[4787]: I1203 17:27:20.713297 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9","Type":"ContainerStarted","Data":"9f241108687bb6e04e230409b1c9e4b4fddf6514f61b231103a7c8dd3186ab26"} Dec 03 17:27:20 crc kubenswrapper[4787]: I1203 17:27:20.714703 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"7da69ee7-38ad-4cce-9bf0-140e7283a775","Type":"ContainerStarted","Data":"fa6844654ff5845c695c87864c6e6f0aadc1bea36b028a8927a246c797080a74"} Dec 03 17:27:20 crc kubenswrapper[4787]: I1203 17:27:20.716983 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" event={"ID":"cb532cad-464c-4fe6-99de-b7897de9ac51","Type":"ContainerStarted","Data":"0a4f976a3548b519694453ddb6988941293f09aff494dc0da69af2c8e50c4761"} Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.763976 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" event={"ID":"4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a","Type":"ContainerStarted","Data":"51df5cbba0ce7d0d520c3e16329d4fc07564499f9172c31feed89912e68dcb0c"} Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.764935 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.765874 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-compactor-0" event={"ID":"db4243ab-4a52-4991-89a6-96e1ef9a348f","Type":"ContainerStarted","Data":"ecb7ccbdc970464d64c3cf0352445029e98cbc669715cdf011b0fe3cd1518575"} Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.765978 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.768205 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" event={"ID":"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc","Type":"ContainerStarted","Data":"23311d1e6e21ae86ded12f6817ad773287bd71b4fec020dcb206ee05c977d79f"} Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.769595 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-index-gateway-0" event={"ID":"b4de518a-6a4d-441b-b04c-d35ca9a8b5f9","Type":"ContainerStarted","Data":"70f535a280db60d1f17fa28593ef1d95f1af47206ee97d3e0e60a9273351fd86"} Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.769851 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.772550 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-ingester-0" event={"ID":"7da69ee7-38ad-4cce-9bf0-140e7283a775","Type":"ContainerStarted","Data":"dea21bb786002421fb3dbcb8650695c5a44cbcae12bc6e03f6fbe94f422aaf01"} Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.772649 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.773747 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" event={"ID":"e1be9303-41d5-431a-bb49-b0e104ce4625","Type":"ContainerStarted","Data":"5cde30453f1dedac52eccef92233dde0eeb8bfc045141944180451fa6c63bb8f"} Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.773876 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.775090 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" event={"ID":"b43db9b8-11f8-4fb7-82d2-6b38e34e1a02","Type":"ContainerStarted","Data":"b72fe5d60831b9bc30ffca359f87c89824361f57d96f1d0d5bcfc3834f76a256"} Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.775327 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.782397 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" podStartSLOduration=1.98350425 podStartE2EDuration="6.782376923s" podCreationTimestamp="2025-12-03 17:27:18 +0000 UTC" firstStartedPulling="2025-12-03 17:27:18.848800787 +0000 UTC m=+855.666271746" lastFinishedPulling="2025-12-03 17:27:23.64767346 +0000 UTC m=+860.465144419" observedRunningTime="2025-12-03 17:27:24.780476563 +0000 UTC m=+861.597947522" watchObservedRunningTime="2025-12-03 17:27:24.782376923 +0000 UTC m=+861.599847882" Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.819711 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-index-gateway-0" podStartSLOduration=3.8489052519999998 podStartE2EDuration="6.819680475s" podCreationTimestamp="2025-12-03 17:27:18 +0000 UTC" firstStartedPulling="2025-12-03 17:27:20.629313654 +0000 UTC m=+857.446784613" lastFinishedPulling="2025-12-03 17:27:23.600088877 +0000 UTC m=+860.417559836" observedRunningTime="2025-12-03 17:27:24.809729303 +0000 UTC m=+861.627200272" watchObservedRunningTime="2025-12-03 17:27:24.819680475 +0000 UTC m=+861.637151434" Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.848237 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-compactor-0" podStartSLOduration=3.779171616 podStartE2EDuration="6.848204626s" podCreationTimestamp="2025-12-03 17:27:18 +0000 UTC" firstStartedPulling="2025-12-03 17:27:20.5919334 +0000 UTC m=+857.409404359" lastFinishedPulling="2025-12-03 17:27:23.66096641 +0000 UTC m=+860.478437369" observedRunningTime="2025-12-03 17:27:24.836140648 +0000 UTC m=+861.653611617" watchObservedRunningTime="2025-12-03 17:27:24.848204626 +0000 UTC m=+861.665675585" Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.864281 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" podStartSLOduration=2.201663464 podStartE2EDuration="6.864254338s" podCreationTimestamp="2025-12-03 17:27:18 +0000 UTC" firstStartedPulling="2025-12-03 17:27:18.997815351 +0000 UTC m=+855.815286310" lastFinishedPulling="2025-12-03 17:27:23.660406225 +0000 UTC m=+860.477877184" observedRunningTime="2025-12-03 17:27:24.857036598 +0000 UTC m=+861.674507567" watchObservedRunningTime="2025-12-03 17:27:24.864254338 +0000 UTC m=+861.681725297" Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.884045 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-ingester-0" podStartSLOduration=3.712292896 podStartE2EDuration="6.883993498s" podCreationTimestamp="2025-12-03 17:27:18 +0000 UTC" firstStartedPulling="2025-12-03 17:27:20.517506551 +0000 UTC m=+857.334977520" lastFinishedPulling="2025-12-03 17:27:23.689207163 +0000 UTC m=+860.506678122" observedRunningTime="2025-12-03 17:27:24.882464658 +0000 UTC m=+861.699935637" watchObservedRunningTime="2025-12-03 17:27:24.883993498 +0000 UTC m=+861.701464457" Dec 03 17:27:24 crc kubenswrapper[4787]: I1203 17:27:24.915457 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" podStartSLOduration=2.204678643 podStartE2EDuration="6.915424286s" podCreationTimestamp="2025-12-03 17:27:18 +0000 UTC" firstStartedPulling="2025-12-03 17:27:18.94235479 +0000 UTC m=+855.759825749" lastFinishedPulling="2025-12-03 17:27:23.653100443 +0000 UTC m=+860.470571392" observedRunningTime="2025-12-03 17:27:24.910709031 +0000 UTC m=+861.728179990" watchObservedRunningTime="2025-12-03 17:27:24.915424286 +0000 UTC m=+861.732895255" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.180646 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-tt67b"] Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.197062 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.204090 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tt67b"] Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.344967 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-catalog-content\") pod \"redhat-marketplace-tt67b\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.345051 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-utilities\") pod \"redhat-marketplace-tt67b\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.345075 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmq7q\" (UniqueName: \"kubernetes.io/projected/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-kube-api-access-qmq7q\") pod \"redhat-marketplace-tt67b\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.446869 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-catalog-content\") pod \"redhat-marketplace-tt67b\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.446939 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-utilities\") pod \"redhat-marketplace-tt67b\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.446962 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmq7q\" (UniqueName: \"kubernetes.io/projected/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-kube-api-access-qmq7q\") pod \"redhat-marketplace-tt67b\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.447525 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-catalog-content\") pod \"redhat-marketplace-tt67b\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.447593 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-utilities\") pod \"redhat-marketplace-tt67b\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.467970 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmq7q\" (UniqueName: \"kubernetes.io/projected/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-kube-api-access-qmq7q\") pod \"redhat-marketplace-tt67b\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.534846 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.803601 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" event={"ID":"35486d7a-dcc2-40bf-b2b0-869ecfe46ccc","Type":"ContainerStarted","Data":"516125648a57f0655446c6f0e2e40efa00ef91db83ad1099a9a1b853ffe7e90b"} Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.804246 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.804368 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.819859 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.831288 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.846257 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-dzw77" podStartSLOduration=3.133619552 podStartE2EDuration="8.84624002s" podCreationTimestamp="2025-12-03 17:27:18 +0000 UTC" firstStartedPulling="2025-12-03 17:27:20.013911642 +0000 UTC m=+856.831382611" lastFinishedPulling="2025-12-03 17:27:25.72653212 +0000 UTC m=+862.544003079" observedRunningTime="2025-12-03 17:27:26.836989066 +0000 UTC m=+863.654460025" watchObservedRunningTime="2025-12-03 17:27:26.84624002 +0000 UTC m=+863.663710979" Dec 03 17:27:26 crc kubenswrapper[4787]: I1203 17:27:26.850591 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-tt67b"] Dec 03 17:27:27 crc kubenswrapper[4787]: I1203 17:27:27.810563 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" event={"ID":"cb532cad-464c-4fe6-99de-b7897de9ac51","Type":"ContainerStarted","Data":"af4bad8382b199805c3478ab9861287d0f2321d30d9a7c1fb947783b9bc86925"} Dec 03 17:27:27 crc kubenswrapper[4787]: I1203 17:27:27.812201 4787 generic.go:334] "Generic (PLEG): container finished" podID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" containerID="6b282555770a36861d5a128d2fc45f214dfd0016e77c63ae00e269502e6b5383" exitCode=0 Dec 03 17:27:27 crc kubenswrapper[4787]: I1203 17:27:27.812276 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tt67b" event={"ID":"e7fc84ba-2781-4bfc-ae0f-cb35c780490e","Type":"ContainerDied","Data":"6b282555770a36861d5a128d2fc45f214dfd0016e77c63ae00e269502e6b5383"} Dec 03 17:27:27 crc kubenswrapper[4787]: I1203 17:27:27.812340 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tt67b" event={"ID":"e7fc84ba-2781-4bfc-ae0f-cb35c780490e","Type":"ContainerStarted","Data":"16aa94d220ae9b8fa7249a1e80f8d43ecc152377f130aa4ef0e01067af61672d"} Dec 03 17:27:28 crc kubenswrapper[4787]: I1203 17:27:28.820818 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" event={"ID":"cb532cad-464c-4fe6-99de-b7897de9ac51","Type":"ContainerStarted","Data":"8d3dcfa228c82b5397e5cbd33192078690ce9f1299f678a01274252b50b05294"} Dec 03 17:27:28 crc kubenswrapper[4787]: I1203 17:27:28.821441 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:28 crc kubenswrapper[4787]: I1203 17:27:28.827331 4787 generic.go:334] "Generic (PLEG): container finished" podID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" containerID="e98a82cb62349536bd9f2d8b706644fae8755e0dda694fcbed89fdafb6771e81" exitCode=0 Dec 03 17:27:28 crc kubenswrapper[4787]: I1203 17:27:28.827419 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tt67b" event={"ID":"e7fc84ba-2781-4bfc-ae0f-cb35c780490e","Type":"ContainerDied","Data":"e98a82cb62349536bd9f2d8b706644fae8755e0dda694fcbed89fdafb6771e81"} Dec 03 17:27:28 crc kubenswrapper[4787]: I1203 17:27:28.831983 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:28 crc kubenswrapper[4787]: I1203 17:27:28.848693 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" podStartSLOduration=3.42115822 podStartE2EDuration="10.848665668s" podCreationTimestamp="2025-12-03 17:27:18 +0000 UTC" firstStartedPulling="2025-12-03 17:27:19.890895893 +0000 UTC m=+856.708366852" lastFinishedPulling="2025-12-03 17:27:27.318403341 +0000 UTC m=+864.135874300" observedRunningTime="2025-12-03 17:27:28.839853656 +0000 UTC m=+865.657324625" watchObservedRunningTime="2025-12-03 17:27:28.848665668 +0000 UTC m=+865.666136647" Dec 03 17:27:29 crc kubenswrapper[4787]: I1203 17:27:29.419856 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:29 crc kubenswrapper[4787]: I1203 17:27:29.430804 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-gateway-57f76f4c5d-v45xd" Dec 03 17:27:29 crc kubenswrapper[4787]: I1203 17:27:29.837733 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tt67b" event={"ID":"e7fc84ba-2781-4bfc-ae0f-cb35c780490e","Type":"ContainerStarted","Data":"1fa0b7fb1957b134d12b534cdd08c17fc0c37a5b483e1bc421321509123680da"} Dec 03 17:27:29 crc kubenswrapper[4787]: I1203 17:27:29.860900 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-tt67b" podStartSLOduration=2.360190589 podStartE2EDuration="3.860881048s" podCreationTimestamp="2025-12-03 17:27:26 +0000 UTC" firstStartedPulling="2025-12-03 17:27:27.813839414 +0000 UTC m=+864.631310373" lastFinishedPulling="2025-12-03 17:27:29.314529873 +0000 UTC m=+866.132000832" observedRunningTime="2025-12-03 17:27:29.855526237 +0000 UTC m=+866.672997216" watchObservedRunningTime="2025-12-03 17:27:29.860881048 +0000 UTC m=+866.678352007" Dec 03 17:27:36 crc kubenswrapper[4787]: I1203 17:27:36.535458 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:36 crc kubenswrapper[4787]: I1203 17:27:36.535827 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:36 crc kubenswrapper[4787]: I1203 17:27:36.620385 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:36 crc kubenswrapper[4787]: I1203 17:27:36.918519 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:36 crc kubenswrapper[4787]: I1203 17:27:36.960176 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tt67b"] Dec 03 17:27:38 crc kubenswrapper[4787]: I1203 17:27:38.409095 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-distributor-76cc67bf56-flghj" Dec 03 17:27:38 crc kubenswrapper[4787]: I1203 17:27:38.550920 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-querier-5895d59bb8-xq92s" Dec 03 17:27:38 crc kubenswrapper[4787]: I1203 17:27:38.654774 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-query-frontend-84558f7c9f-zsjlj" Dec 03 17:27:38 crc kubenswrapper[4787]: I1203 17:27:38.889453 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-tt67b" podUID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" containerName="registry-server" containerID="cri-o://1fa0b7fb1957b134d12b534cdd08c17fc0c37a5b483e1bc421321509123680da" gracePeriod=2 Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.264665 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-db5gf"] Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.266351 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.274256 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-db5gf"] Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.356813 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-utilities\") pod \"certified-operators-db5gf\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.356877 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-catalog-content\") pod \"certified-operators-db5gf\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.356931 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2kzbd\" (UniqueName: \"kubernetes.io/projected/56c3a374-1c38-4c2d-8573-d84c9af99538-kube-api-access-2kzbd\") pod \"certified-operators-db5gf\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.459042 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2kzbd\" (UniqueName: \"kubernetes.io/projected/56c3a374-1c38-4c2d-8573-d84c9af99538-kube-api-access-2kzbd\") pod \"certified-operators-db5gf\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.459190 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-utilities\") pod \"certified-operators-db5gf\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.459228 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-catalog-content\") pod \"certified-operators-db5gf\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.459897 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-utilities\") pod \"certified-operators-db5gf\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.459940 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-catalog-content\") pod \"certified-operators-db5gf\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.481926 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2kzbd\" (UniqueName: \"kubernetes.io/projected/56c3a374-1c38-4c2d-8573-d84c9af99538-kube-api-access-2kzbd\") pod \"certified-operators-db5gf\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.583203 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.900639 4787 generic.go:334] "Generic (PLEG): container finished" podID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" containerID="1fa0b7fb1957b134d12b534cdd08c17fc0c37a5b483e1bc421321509123680da" exitCode=0 Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.901006 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tt67b" event={"ID":"e7fc84ba-2781-4bfc-ae0f-cb35c780490e","Type":"ContainerDied","Data":"1fa0b7fb1957b134d12b534cdd08c17fc0c37a5b483e1bc421321509123680da"} Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.921455 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-index-gateway-0" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.922375 4787 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.922427 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="7da69ee7-38ad-4cce-9bf0-140e7283a775" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.931753 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-compactor-0" Dec 03 17:27:39 crc kubenswrapper[4787]: I1203 17:27:39.977414 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.032523 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-db5gf"] Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.071904 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmq7q\" (UniqueName: \"kubernetes.io/projected/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-kube-api-access-qmq7q\") pod \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.071991 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-catalog-content\") pod \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.072119 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-utilities\") pod \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\" (UID: \"e7fc84ba-2781-4bfc-ae0f-cb35c780490e\") " Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.076869 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-utilities" (OuterVolumeSpecName: "utilities") pod "e7fc84ba-2781-4bfc-ae0f-cb35c780490e" (UID: "e7fc84ba-2781-4bfc-ae0f-cb35c780490e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.105313 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-kube-api-access-qmq7q" (OuterVolumeSpecName: "kube-api-access-qmq7q") pod "e7fc84ba-2781-4bfc-ae0f-cb35c780490e" (UID: "e7fc84ba-2781-4bfc-ae0f-cb35c780490e"). InnerVolumeSpecName "kube-api-access-qmq7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.117443 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e7fc84ba-2781-4bfc-ae0f-cb35c780490e" (UID: "e7fc84ba-2781-4bfc-ae0f-cb35c780490e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.173741 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmq7q\" (UniqueName: \"kubernetes.io/projected/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-kube-api-access-qmq7q\") on node \"crc\" DevicePath \"\"" Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.173932 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.174036 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e7fc84ba-2781-4bfc-ae0f-cb35c780490e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.910363 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-tt67b" event={"ID":"e7fc84ba-2781-4bfc-ae0f-cb35c780490e","Type":"ContainerDied","Data":"16aa94d220ae9b8fa7249a1e80f8d43ecc152377f130aa4ef0e01067af61672d"} Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.910660 4787 scope.go:117] "RemoveContainer" containerID="1fa0b7fb1957b134d12b534cdd08c17fc0c37a5b483e1bc421321509123680da" Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.910416 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-tt67b" Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.912354 4787 generic.go:334] "Generic (PLEG): container finished" podID="56c3a374-1c38-4c2d-8573-d84c9af99538" containerID="7b6a43594c03c3e12c911bf28a18a72038dfb451de5772ebfabe1697e9ef44af" exitCode=0 Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.912395 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db5gf" event={"ID":"56c3a374-1c38-4c2d-8573-d84c9af99538","Type":"ContainerDied","Data":"7b6a43594c03c3e12c911bf28a18a72038dfb451de5772ebfabe1697e9ef44af"} Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.912426 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db5gf" event={"ID":"56c3a374-1c38-4c2d-8573-d84c9af99538","Type":"ContainerStarted","Data":"e371fafabb9e14976d3d0b90f069f4cd8f2d205c2fc0ba9d4c1b8b6589cf1dab"} Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.930000 4787 scope.go:117] "RemoveContainer" containerID="e98a82cb62349536bd9f2d8b706644fae8755e0dda694fcbed89fdafb6771e81" Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.970820 4787 scope.go:117] "RemoveContainer" containerID="6b282555770a36861d5a128d2fc45f214dfd0016e77c63ae00e269502e6b5383" Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.979113 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-tt67b"] Dec 03 17:27:40 crc kubenswrapper[4787]: I1203 17:27:40.984602 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-tt67b"] Dec 03 17:27:41 crc kubenswrapper[4787]: I1203 17:27:41.778452 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" path="/var/lib/kubelet/pods/e7fc84ba-2781-4bfc-ae0f-cb35c780490e/volumes" Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.861927 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ps7lt"] Dec 03 17:27:42 crc kubenswrapper[4787]: E1203 17:27:42.862517 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" containerName="extract-content" Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.862531 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" containerName="extract-content" Dec 03 17:27:42 crc kubenswrapper[4787]: E1203 17:27:42.862544 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" containerName="registry-server" Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.862551 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" containerName="registry-server" Dec 03 17:27:42 crc kubenswrapper[4787]: E1203 17:27:42.862568 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" containerName="extract-utilities" Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.862577 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" containerName="extract-utilities" Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.862701 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7fc84ba-2781-4bfc-ae0f-cb35c780490e" containerName="registry-server" Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.863657 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.872629 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ps7lt"] Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.919878 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-catalog-content\") pod \"community-operators-ps7lt\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.919968 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2krr4\" (UniqueName: \"kubernetes.io/projected/e78f781b-ed98-4fe2-bd67-f14012241c14-kube-api-access-2krr4\") pod \"community-operators-ps7lt\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.920049 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-utilities\") pod \"community-operators-ps7lt\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.931684 4787 generic.go:334] "Generic (PLEG): container finished" podID="56c3a374-1c38-4c2d-8573-d84c9af99538" containerID="25fc19c4ae71ecfd3f086f8fd3849c8425028d7fa26f4025baa2eaa452a14854" exitCode=0 Dec 03 17:27:42 crc kubenswrapper[4787]: I1203 17:27:42.931727 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db5gf" event={"ID":"56c3a374-1c38-4c2d-8573-d84c9af99538","Type":"ContainerDied","Data":"25fc19c4ae71ecfd3f086f8fd3849c8425028d7fa26f4025baa2eaa452a14854"} Dec 03 17:27:43 crc kubenswrapper[4787]: I1203 17:27:43.021532 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-catalog-content\") pod \"community-operators-ps7lt\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:43 crc kubenswrapper[4787]: I1203 17:27:43.021633 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2krr4\" (UniqueName: \"kubernetes.io/projected/e78f781b-ed98-4fe2-bd67-f14012241c14-kube-api-access-2krr4\") pod \"community-operators-ps7lt\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:43 crc kubenswrapper[4787]: I1203 17:27:43.021675 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-utilities\") pod \"community-operators-ps7lt\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:43 crc kubenswrapper[4787]: I1203 17:27:43.022231 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-catalog-content\") pod \"community-operators-ps7lt\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:43 crc kubenswrapper[4787]: I1203 17:27:43.022262 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-utilities\") pod \"community-operators-ps7lt\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:43 crc kubenswrapper[4787]: I1203 17:27:43.053218 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2krr4\" (UniqueName: \"kubernetes.io/projected/e78f781b-ed98-4fe2-bd67-f14012241c14-kube-api-access-2krr4\") pod \"community-operators-ps7lt\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:43 crc kubenswrapper[4787]: I1203 17:27:43.189031 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:43 crc kubenswrapper[4787]: I1203 17:27:43.983920 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ps7lt"] Dec 03 17:27:44 crc kubenswrapper[4787]: I1203 17:27:44.951117 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ps7lt" event={"ID":"e78f781b-ed98-4fe2-bd67-f14012241c14","Type":"ContainerStarted","Data":"1210fe83e068e61201a1fa8bee0cc208079b1c83f117d0b4080d0b64f461ff59"} Dec 03 17:27:45 crc kubenswrapper[4787]: I1203 17:27:45.959049 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db5gf" event={"ID":"56c3a374-1c38-4c2d-8573-d84c9af99538","Type":"ContainerStarted","Data":"e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91"} Dec 03 17:27:45 crc kubenswrapper[4787]: I1203 17:27:45.960576 4787 generic.go:334] "Generic (PLEG): container finished" podID="e78f781b-ed98-4fe2-bd67-f14012241c14" containerID="385ae482ba889f171416e969ac2dcb99bdf134b0585f53a5164c197a778984b0" exitCode=0 Dec 03 17:27:45 crc kubenswrapper[4787]: I1203 17:27:45.960619 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ps7lt" event={"ID":"e78f781b-ed98-4fe2-bd67-f14012241c14","Type":"ContainerDied","Data":"385ae482ba889f171416e969ac2dcb99bdf134b0585f53a5164c197a778984b0"} Dec 03 17:27:45 crc kubenswrapper[4787]: I1203 17:27:45.979581 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-db5gf" podStartSLOduration=4.185023639 podStartE2EDuration="6.979561241s" podCreationTimestamp="2025-12-03 17:27:39 +0000 UTC" firstStartedPulling="2025-12-03 17:27:40.913777523 +0000 UTC m=+877.731248482" lastFinishedPulling="2025-12-03 17:27:43.708315125 +0000 UTC m=+880.525786084" observedRunningTime="2025-12-03 17:27:45.976460479 +0000 UTC m=+882.793931448" watchObservedRunningTime="2025-12-03 17:27:45.979561241 +0000 UTC m=+882.797032200" Dec 03 17:27:46 crc kubenswrapper[4787]: I1203 17:27:46.969207 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ps7lt" event={"ID":"e78f781b-ed98-4fe2-bd67-f14012241c14","Type":"ContainerStarted","Data":"e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5"} Dec 03 17:27:48 crc kubenswrapper[4787]: I1203 17:27:48.982955 4787 generic.go:334] "Generic (PLEG): container finished" podID="e78f781b-ed98-4fe2-bd67-f14012241c14" containerID="e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5" exitCode=0 Dec 03 17:27:48 crc kubenswrapper[4787]: I1203 17:27:48.983006 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ps7lt" event={"ID":"e78f781b-ed98-4fe2-bd67-f14012241c14","Type":"ContainerDied","Data":"e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5"} Dec 03 17:27:49 crc kubenswrapper[4787]: I1203 17:27:49.584044 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:49 crc kubenswrapper[4787]: I1203 17:27:49.584383 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:49 crc kubenswrapper[4787]: I1203 17:27:49.636052 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:49 crc kubenswrapper[4787]: I1203 17:27:49.914117 4787 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: this instance owns no tokens Dec 03 17:27:49 crc kubenswrapper[4787]: I1203 17:27:49.914192 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="7da69ee7-38ad-4cce-9bf0-140e7283a775" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 17:27:49 crc kubenswrapper[4787]: I1203 17:27:49.995262 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ps7lt" event={"ID":"e78f781b-ed98-4fe2-bd67-f14012241c14","Type":"ContainerStarted","Data":"db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f"} Dec 03 17:27:50 crc kubenswrapper[4787]: I1203 17:27:50.016597 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ps7lt" podStartSLOduration=4.481720251 podStartE2EDuration="8.016575104s" podCreationTimestamp="2025-12-03 17:27:42 +0000 UTC" firstStartedPulling="2025-12-03 17:27:45.961979208 +0000 UTC m=+882.779450167" lastFinishedPulling="2025-12-03 17:27:49.496834061 +0000 UTC m=+886.314305020" observedRunningTime="2025-12-03 17:27:50.013504994 +0000 UTC m=+886.830975953" watchObservedRunningTime="2025-12-03 17:27:50.016575104 +0000 UTC m=+886.834046063" Dec 03 17:27:53 crc kubenswrapper[4787]: I1203 17:27:53.190210 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:53 crc kubenswrapper[4787]: I1203 17:27:53.192116 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:53 crc kubenswrapper[4787]: I1203 17:27:53.229238 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:54 crc kubenswrapper[4787]: I1203 17:27:54.057145 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:54 crc kubenswrapper[4787]: I1203 17:27:54.658222 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ps7lt"] Dec 03 17:27:56 crc kubenswrapper[4787]: I1203 17:27:56.033940 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ps7lt" podUID="e78f781b-ed98-4fe2-bd67-f14012241c14" containerName="registry-server" containerID="cri-o://db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f" gracePeriod=2 Dec 03 17:27:56 crc kubenswrapper[4787]: I1203 17:27:56.409897 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:56 crc kubenswrapper[4787]: I1203 17:27:56.499909 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-catalog-content\") pod \"e78f781b-ed98-4fe2-bd67-f14012241c14\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " Dec 03 17:27:56 crc kubenswrapper[4787]: I1203 17:27:56.499959 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-utilities\") pod \"e78f781b-ed98-4fe2-bd67-f14012241c14\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " Dec 03 17:27:56 crc kubenswrapper[4787]: I1203 17:27:56.500152 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2krr4\" (UniqueName: \"kubernetes.io/projected/e78f781b-ed98-4fe2-bd67-f14012241c14-kube-api-access-2krr4\") pod \"e78f781b-ed98-4fe2-bd67-f14012241c14\" (UID: \"e78f781b-ed98-4fe2-bd67-f14012241c14\") " Dec 03 17:27:56 crc kubenswrapper[4787]: I1203 17:27:56.501339 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-utilities" (OuterVolumeSpecName: "utilities") pod "e78f781b-ed98-4fe2-bd67-f14012241c14" (UID: "e78f781b-ed98-4fe2-bd67-f14012241c14"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:27:56 crc kubenswrapper[4787]: I1203 17:27:56.505542 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e78f781b-ed98-4fe2-bd67-f14012241c14-kube-api-access-2krr4" (OuterVolumeSpecName: "kube-api-access-2krr4") pod "e78f781b-ed98-4fe2-bd67-f14012241c14" (UID: "e78f781b-ed98-4fe2-bd67-f14012241c14"). InnerVolumeSpecName "kube-api-access-2krr4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:27:56 crc kubenswrapper[4787]: I1203 17:27:56.547006 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e78f781b-ed98-4fe2-bd67-f14012241c14" (UID: "e78f781b-ed98-4fe2-bd67-f14012241c14"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:27:56 crc kubenswrapper[4787]: I1203 17:27:56.600874 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:27:56 crc kubenswrapper[4787]: I1203 17:27:56.601283 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e78f781b-ed98-4fe2-bd67-f14012241c14-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:27:56 crc kubenswrapper[4787]: I1203 17:27:56.601303 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2krr4\" (UniqueName: \"kubernetes.io/projected/e78f781b-ed98-4fe2-bd67-f14012241c14-kube-api-access-2krr4\") on node \"crc\" DevicePath \"\"" Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.043373 4787 generic.go:334] "Generic (PLEG): container finished" podID="e78f781b-ed98-4fe2-bd67-f14012241c14" containerID="db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f" exitCode=0 Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.043412 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ps7lt" event={"ID":"e78f781b-ed98-4fe2-bd67-f14012241c14","Type":"ContainerDied","Data":"db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f"} Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.043440 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ps7lt" event={"ID":"e78f781b-ed98-4fe2-bd67-f14012241c14","Type":"ContainerDied","Data":"1210fe83e068e61201a1fa8bee0cc208079b1c83f117d0b4080d0b64f461ff59"} Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.043459 4787 scope.go:117] "RemoveContainer" containerID="db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f" Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.043593 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ps7lt" Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.068143 4787 scope.go:117] "RemoveContainer" containerID="e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5" Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.076317 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ps7lt"] Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.081296 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ps7lt"] Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.103401 4787 scope.go:117] "RemoveContainer" containerID="385ae482ba889f171416e969ac2dcb99bdf134b0585f53a5164c197a778984b0" Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.116415 4787 scope.go:117] "RemoveContainer" containerID="db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f" Dec 03 17:27:57 crc kubenswrapper[4787]: E1203 17:27:57.116848 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f\": container with ID starting with db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f not found: ID does not exist" containerID="db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f" Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.116886 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f"} err="failed to get container status \"db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f\": rpc error: code = NotFound desc = could not find container \"db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f\": container with ID starting with db2f390636795e4bd42a323340282c3c0de4654e751792555f5c8968d8100b8f not found: ID does not exist" Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.116909 4787 scope.go:117] "RemoveContainer" containerID="e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5" Dec 03 17:27:57 crc kubenswrapper[4787]: E1203 17:27:57.117239 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5\": container with ID starting with e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5 not found: ID does not exist" containerID="e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5" Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.117283 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5"} err="failed to get container status \"e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5\": rpc error: code = NotFound desc = could not find container \"e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5\": container with ID starting with e91559cab643cdc93ef1305e702f1a4cf18833ec3561fff348dd34e15d4490c5 not found: ID does not exist" Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.117315 4787 scope.go:117] "RemoveContainer" containerID="385ae482ba889f171416e969ac2dcb99bdf134b0585f53a5164c197a778984b0" Dec 03 17:27:57 crc kubenswrapper[4787]: E1203 17:27:57.117582 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"385ae482ba889f171416e969ac2dcb99bdf134b0585f53a5164c197a778984b0\": container with ID starting with 385ae482ba889f171416e969ac2dcb99bdf134b0585f53a5164c197a778984b0 not found: ID does not exist" containerID="385ae482ba889f171416e969ac2dcb99bdf134b0585f53a5164c197a778984b0" Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.117608 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"385ae482ba889f171416e969ac2dcb99bdf134b0585f53a5164c197a778984b0"} err="failed to get container status \"385ae482ba889f171416e969ac2dcb99bdf134b0585f53a5164c197a778984b0\": rpc error: code = NotFound desc = could not find container \"385ae482ba889f171416e969ac2dcb99bdf134b0585f53a5164c197a778984b0\": container with ID starting with 385ae482ba889f171416e969ac2dcb99bdf134b0585f53a5164c197a778984b0 not found: ID does not exist" Dec 03 17:27:57 crc kubenswrapper[4787]: I1203 17:27:57.774058 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e78f781b-ed98-4fe2-bd67-f14012241c14" path="/var/lib/kubelet/pods/e78f781b-ed98-4fe2-bd67-f14012241c14/volumes" Dec 03 17:27:59 crc kubenswrapper[4787]: I1203 17:27:59.645395 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:27:59 crc kubenswrapper[4787]: I1203 17:27:59.910581 4787 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 03 17:27:59 crc kubenswrapper[4787]: I1203 17:27:59.910644 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="7da69ee7-38ad-4cce-9bf0-140e7283a775" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.064249 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-db5gf"] Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.067811 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-db5gf" podUID="56c3a374-1c38-4c2d-8573-d84c9af99538" containerName="registry-server" containerID="cri-o://e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91" gracePeriod=2 Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.542970 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.655163 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2kzbd\" (UniqueName: \"kubernetes.io/projected/56c3a374-1c38-4c2d-8573-d84c9af99538-kube-api-access-2kzbd\") pod \"56c3a374-1c38-4c2d-8573-d84c9af99538\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.655306 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-catalog-content\") pod \"56c3a374-1c38-4c2d-8573-d84c9af99538\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.655337 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-utilities\") pod \"56c3a374-1c38-4c2d-8573-d84c9af99538\" (UID: \"56c3a374-1c38-4c2d-8573-d84c9af99538\") " Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.656310 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-utilities" (OuterVolumeSpecName: "utilities") pod "56c3a374-1c38-4c2d-8573-d84c9af99538" (UID: "56c3a374-1c38-4c2d-8573-d84c9af99538"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.661493 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56c3a374-1c38-4c2d-8573-d84c9af99538-kube-api-access-2kzbd" (OuterVolumeSpecName: "kube-api-access-2kzbd") pod "56c3a374-1c38-4c2d-8573-d84c9af99538" (UID: "56c3a374-1c38-4c2d-8573-d84c9af99538"). InnerVolumeSpecName "kube-api-access-2kzbd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.728576 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56c3a374-1c38-4c2d-8573-d84c9af99538" (UID: "56c3a374-1c38-4c2d-8573-d84c9af99538"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.757002 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2kzbd\" (UniqueName: \"kubernetes.io/projected/56c3a374-1c38-4c2d-8573-d84c9af99538-kube-api-access-2kzbd\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.757078 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:00 crc kubenswrapper[4787]: I1203 17:28:00.757091 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56c3a374-1c38-4c2d-8573-d84c9af99538-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.076108 4787 generic.go:334] "Generic (PLEG): container finished" podID="56c3a374-1c38-4c2d-8573-d84c9af99538" containerID="e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91" exitCode=0 Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.076147 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db5gf" event={"ID":"56c3a374-1c38-4c2d-8573-d84c9af99538","Type":"ContainerDied","Data":"e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91"} Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.076174 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-db5gf" event={"ID":"56c3a374-1c38-4c2d-8573-d84c9af99538","Type":"ContainerDied","Data":"e371fafabb9e14976d3d0b90f069f4cd8f2d205c2fc0ba9d4c1b8b6589cf1dab"} Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.076214 4787 scope.go:117] "RemoveContainer" containerID="e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91" Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.076592 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-db5gf" Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.105131 4787 scope.go:117] "RemoveContainer" containerID="25fc19c4ae71ecfd3f086f8fd3849c8425028d7fa26f4025baa2eaa452a14854" Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.109446 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-db5gf"] Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.114121 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-db5gf"] Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.141503 4787 scope.go:117] "RemoveContainer" containerID="7b6a43594c03c3e12c911bf28a18a72038dfb451de5772ebfabe1697e9ef44af" Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.166812 4787 scope.go:117] "RemoveContainer" containerID="e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91" Dec 03 17:28:01 crc kubenswrapper[4787]: E1203 17:28:01.167618 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91\": container with ID starting with e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91 not found: ID does not exist" containerID="e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91" Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.167689 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91"} err="failed to get container status \"e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91\": rpc error: code = NotFound desc = could not find container \"e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91\": container with ID starting with e52c2a8c024c2e2741d64427ac56c939171b414465a565c04345094b1c741e91 not found: ID does not exist" Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.167720 4787 scope.go:117] "RemoveContainer" containerID="25fc19c4ae71ecfd3f086f8fd3849c8425028d7fa26f4025baa2eaa452a14854" Dec 03 17:28:01 crc kubenswrapper[4787]: E1203 17:28:01.168094 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25fc19c4ae71ecfd3f086f8fd3849c8425028d7fa26f4025baa2eaa452a14854\": container with ID starting with 25fc19c4ae71ecfd3f086f8fd3849c8425028d7fa26f4025baa2eaa452a14854 not found: ID does not exist" containerID="25fc19c4ae71ecfd3f086f8fd3849c8425028d7fa26f4025baa2eaa452a14854" Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.168123 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25fc19c4ae71ecfd3f086f8fd3849c8425028d7fa26f4025baa2eaa452a14854"} err="failed to get container status \"25fc19c4ae71ecfd3f086f8fd3849c8425028d7fa26f4025baa2eaa452a14854\": rpc error: code = NotFound desc = could not find container \"25fc19c4ae71ecfd3f086f8fd3849c8425028d7fa26f4025baa2eaa452a14854\": container with ID starting with 25fc19c4ae71ecfd3f086f8fd3849c8425028d7fa26f4025baa2eaa452a14854 not found: ID does not exist" Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.168141 4787 scope.go:117] "RemoveContainer" containerID="7b6a43594c03c3e12c911bf28a18a72038dfb451de5772ebfabe1697e9ef44af" Dec 03 17:28:01 crc kubenswrapper[4787]: E1203 17:28:01.168392 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b6a43594c03c3e12c911bf28a18a72038dfb451de5772ebfabe1697e9ef44af\": container with ID starting with 7b6a43594c03c3e12c911bf28a18a72038dfb451de5772ebfabe1697e9ef44af not found: ID does not exist" containerID="7b6a43594c03c3e12c911bf28a18a72038dfb451de5772ebfabe1697e9ef44af" Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.168412 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b6a43594c03c3e12c911bf28a18a72038dfb451de5772ebfabe1697e9ef44af"} err="failed to get container status \"7b6a43594c03c3e12c911bf28a18a72038dfb451de5772ebfabe1697e9ef44af\": rpc error: code = NotFound desc = could not find container \"7b6a43594c03c3e12c911bf28a18a72038dfb451de5772ebfabe1697e9ef44af\": container with ID starting with 7b6a43594c03c3e12c911bf28a18a72038dfb451de5772ebfabe1697e9ef44af not found: ID does not exist" Dec 03 17:28:01 crc kubenswrapper[4787]: I1203 17:28:01.775040 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56c3a374-1c38-4c2d-8573-d84c9af99538" path="/var/lib/kubelet/pods/56c3a374-1c38-4c2d-8573-d84c9af99538/volumes" Dec 03 17:28:09 crc kubenswrapper[4787]: I1203 17:28:09.909078 4787 patch_prober.go:28] interesting pod/logging-loki-ingester-0 container/loki-ingester namespace/openshift-logging: Readiness probe status=failure output="HTTP probe failed with statuscode: 503" start-of-body=Ingester not ready: waiting for 15s after being ready Dec 03 17:28:09 crc kubenswrapper[4787]: I1203 17:28:09.909649 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-logging/logging-loki-ingester-0" podUID="7da69ee7-38ad-4cce-9bf0-140e7283a775" containerName="loki-ingester" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 03 17:28:19 crc kubenswrapper[4787]: I1203 17:28:19.910503 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-logging/logging-loki-ingester-0" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.667990 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-hgqrx"] Dec 03 17:28:39 crc kubenswrapper[4787]: E1203 17:28:39.669054 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56c3a374-1c38-4c2d-8573-d84c9af99538" containerName="registry-server" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.669071 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="56c3a374-1c38-4c2d-8573-d84c9af99538" containerName="registry-server" Dec 03 17:28:39 crc kubenswrapper[4787]: E1203 17:28:39.669109 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56c3a374-1c38-4c2d-8573-d84c9af99538" containerName="extract-utilities" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.669116 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="56c3a374-1c38-4c2d-8573-d84c9af99538" containerName="extract-utilities" Dec 03 17:28:39 crc kubenswrapper[4787]: E1203 17:28:39.669131 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78f781b-ed98-4fe2-bd67-f14012241c14" containerName="extract-content" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.669137 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78f781b-ed98-4fe2-bd67-f14012241c14" containerName="extract-content" Dec 03 17:28:39 crc kubenswrapper[4787]: E1203 17:28:39.669153 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78f781b-ed98-4fe2-bd67-f14012241c14" containerName="extract-utilities" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.669159 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78f781b-ed98-4fe2-bd67-f14012241c14" containerName="extract-utilities" Dec 03 17:28:39 crc kubenswrapper[4787]: E1203 17:28:39.669172 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56c3a374-1c38-4c2d-8573-d84c9af99538" containerName="extract-content" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.669179 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="56c3a374-1c38-4c2d-8573-d84c9af99538" containerName="extract-content" Dec 03 17:28:39 crc kubenswrapper[4787]: E1203 17:28:39.669189 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e78f781b-ed98-4fe2-bd67-f14012241c14" containerName="registry-server" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.669194 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e78f781b-ed98-4fe2-bd67-f14012241c14" containerName="registry-server" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.669308 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="56c3a374-1c38-4c2d-8573-d84c9af99538" containerName="registry-server" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.669321 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e78f781b-ed98-4fe2-bd67-f14012241c14" containerName="registry-server" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.670062 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.672085 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.672778 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.673114 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-kc57g" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.673438 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.674710 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.681319 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.698556 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-hgqrx"] Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.841011 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-metrics\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.841075 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-entrypoint\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.841105 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-token\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.841131 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config-openshift-service-cacrt\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.841165 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-tmp\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.841959 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-hgqrx"] Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.842589 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-trusted-ca\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.842684 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gf2cl\" (UniqueName: \"kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-kube-api-access-gf2cl\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.842781 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-syslog-receiver\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: E1203 17:28:39.842802 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[collector-syslog-receiver collector-token config config-openshift-service-cacrt datadir entrypoint kube-api-access-gf2cl metrics sa-token tmp trusted-ca], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openshift-logging/collector-hgqrx" podUID="55e1ef99-effa-49bf-85ec-1d9cb3ba9229" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.842909 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-datadir\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.842971 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.843140 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-sa-token\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.945216 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-tmp\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.945302 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-trusted-ca\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.946189 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gf2cl\" (UniqueName: \"kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-kube-api-access-gf2cl\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.946250 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-syslog-receiver\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.946278 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-datadir\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.946528 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-datadir\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.946761 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.946849 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-sa-token\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.946908 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-metrics\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.946979 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-entrypoint\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.947054 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-token\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.949635 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config-openshift-service-cacrt\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.947576 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.947581 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-trusted-ca\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.948005 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-entrypoint\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: E1203 17:28:39.947212 4787 secret.go:188] Couldn't get secret openshift-logging/collector-metrics: secret "collector-metrics" not found Dec 03 17:28:39 crc kubenswrapper[4787]: E1203 17:28:39.950050 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-metrics podName:55e1ef99-effa-49bf-85ec-1d9cb3ba9229 nodeName:}" failed. No retries permitted until 2025-12-03 17:28:40.449993066 +0000 UTC m=+937.267464035 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics" (UniqueName: "kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-metrics") pod "collector-hgqrx" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229") : secret "collector-metrics" not found Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.950883 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config-openshift-service-cacrt\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.951797 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-tmp\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.953932 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-syslog-receiver\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.954851 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-token\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.977789 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-sa-token\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:39 crc kubenswrapper[4787]: I1203 17:28:39.991530 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gf2cl\" (UniqueName: \"kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-kube-api-access-gf2cl\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.325583 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-hgqrx" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.338467 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-hgqrx" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.356566 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config-openshift-service-cacrt\") pod \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.357054 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf2cl\" (UniqueName: \"kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-kube-api-access-gf2cl\") pod \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.357206 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-sa-token\") pod \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.357340 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-entrypoint\") pod \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.357423 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-trusted-ca\") pod \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.357094 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config-openshift-service-cacrt" (OuterVolumeSpecName: "config-openshift-service-cacrt") pod "55e1ef99-effa-49bf-85ec-1d9cb3ba9229" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229"). InnerVolumeSpecName "config-openshift-service-cacrt". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.357501 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-token\") pod \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.357654 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-datadir\") pod \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.358255 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-syslog-receiver\") pod \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.357745 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-datadir" (OuterVolumeSpecName: "datadir") pod "55e1ef99-effa-49bf-85ec-1d9cb3ba9229" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229"). InnerVolumeSpecName "datadir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.358124 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-entrypoint" (OuterVolumeSpecName: "entrypoint") pod "55e1ef99-effa-49bf-85ec-1d9cb3ba9229" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229"). InnerVolumeSpecName "entrypoint". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.358171 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "55e1ef99-effa-49bf-85ec-1d9cb3ba9229" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.358344 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-tmp\") pod \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.358375 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config\") pod \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.359228 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config" (OuterVolumeSpecName: "config") pod "55e1ef99-effa-49bf-85ec-1d9cb3ba9229" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.359321 4787 reconciler_common.go:293] "Volume detached for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-entrypoint\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.359343 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.359357 4787 reconciler_common.go:293] "Volume detached for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-datadir\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.359373 4787 reconciler_common.go:293] "Volume detached for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config-openshift-service-cacrt\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.364035 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-sa-token" (OuterVolumeSpecName: "sa-token") pod "55e1ef99-effa-49bf-85ec-1d9cb3ba9229" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229"). InnerVolumeSpecName "sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.364170 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-syslog-receiver" (OuterVolumeSpecName: "collector-syslog-receiver") pod "55e1ef99-effa-49bf-85ec-1d9cb3ba9229" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229"). InnerVolumeSpecName "collector-syslog-receiver". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.364498 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-tmp" (OuterVolumeSpecName: "tmp") pod "55e1ef99-effa-49bf-85ec-1d9cb3ba9229" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229"). InnerVolumeSpecName "tmp". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.365509 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-token" (OuterVolumeSpecName: "collector-token") pod "55e1ef99-effa-49bf-85ec-1d9cb3ba9229" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229"). InnerVolumeSpecName "collector-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.370293 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-kube-api-access-gf2cl" (OuterVolumeSpecName: "kube-api-access-gf2cl") pod "55e1ef99-effa-49bf-85ec-1d9cb3ba9229" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229"). InnerVolumeSpecName "kube-api-access-gf2cl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.460981 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-metrics\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.461208 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf2cl\" (UniqueName: \"kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-kube-api-access-gf2cl\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.461226 4787 reconciler_common.go:293] "Volume detached for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-sa-token\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.461241 4787 reconciler_common.go:293] "Volume detached for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-token\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.461253 4787 reconciler_common.go:293] "Volume detached for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-collector-syslog-receiver\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.461262 4787 reconciler_common.go:293] "Volume detached for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-tmp\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.461270 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.466602 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-metrics\") pod \"collector-hgqrx\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " pod="openshift-logging/collector-hgqrx" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.562778 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-metrics\") pod \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\" (UID: \"55e1ef99-effa-49bf-85ec-1d9cb3ba9229\") " Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.565898 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-metrics" (OuterVolumeSpecName: "metrics") pod "55e1ef99-effa-49bf-85ec-1d9cb3ba9229" (UID: "55e1ef99-effa-49bf-85ec-1d9cb3ba9229"). InnerVolumeSpecName "metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:28:40 crc kubenswrapper[4787]: I1203 17:28:40.665515 4787 reconciler_common.go:293] "Volume detached for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/55e1ef99-effa-49bf-85ec-1d9cb3ba9229-metrics\") on node \"crc\" DevicePath \"\"" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.334882 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-hgqrx" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.405529 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-logging/collector-hgqrx"] Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.412535 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-logging/collector-hgqrx"] Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.433003 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-logging/collector-ghzsk"] Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.439161 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.444523 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-token" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.445035 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-config" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.445337 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-dockercfg-kc57g" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.445476 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-syslog-receiver" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.445834 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-logging"/"collector-metrics" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.456655 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-logging"/"collector-trustbundle" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.471130 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-ghzsk"] Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.595400 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-config\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.595463 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-config-openshift-service-cacrt\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.595494 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-entrypoint\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.595542 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-collector-token\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.595558 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-sa-token\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.595574 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtx6j\" (UniqueName: \"kubernetes.io/projected/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-kube-api-access-qtx6j\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.595598 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-metrics\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.595614 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-trusted-ca\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.595667 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-datadir\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.595687 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-tmp\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.595729 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-collector-syslog-receiver\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.697057 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-datadir\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.697138 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-tmp\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.697197 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-collector-syslog-receiver\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.697254 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-config\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.697288 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-config-openshift-service-cacrt\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.697328 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-entrypoint\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.697378 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-sa-token\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.697401 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-collector-token\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.697426 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtx6j\" (UniqueName: \"kubernetes.io/projected/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-kube-api-access-qtx6j\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.697461 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-metrics\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.697481 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-trusted-ca\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.698108 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"datadir\" (UniqueName: \"kubernetes.io/host-path/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-datadir\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.699903 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-openshift-service-cacrt\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-config-openshift-service-cacrt\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.700198 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"entrypoint\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-entrypoint\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.700832 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-config\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.703336 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-syslog-receiver\" (UniqueName: \"kubernetes.io/secret/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-collector-syslog-receiver\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.703676 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-trusted-ca\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.705913 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"collector-token\" (UniqueName: \"kubernetes.io/secret/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-collector-token\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.709836 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/secret/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-metrics\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.714877 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-tmp\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.725954 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtx6j\" (UniqueName: \"kubernetes.io/projected/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-kube-api-access-qtx6j\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.728118 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sa-token\" (UniqueName: \"kubernetes.io/projected/44f70e1d-0f76-4acb-bf6b-1a5ca89469ca-sa-token\") pod \"collector-ghzsk\" (UID: \"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca\") " pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.770217 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-logging/collector-ghzsk" Dec 03 17:28:41 crc kubenswrapper[4787]: I1203 17:28:41.777210 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="55e1ef99-effa-49bf-85ec-1d9cb3ba9229" path="/var/lib/kubelet/pods/55e1ef99-effa-49bf-85ec-1d9cb3ba9229/volumes" Dec 03 17:28:42 crc kubenswrapper[4787]: I1203 17:28:42.040249 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-logging/collector-ghzsk"] Dec 03 17:28:42 crc kubenswrapper[4787]: I1203 17:28:42.342937 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-ghzsk" event={"ID":"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca","Type":"ContainerStarted","Data":"d7c30d505393bf8ac68eb359387938f870b72620c07fd4e7e4589958fb555305"} Dec 03 17:28:50 crc kubenswrapper[4787]: I1203 17:28:50.432806 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-logging/collector-ghzsk" event={"ID":"44f70e1d-0f76-4acb-bf6b-1a5ca89469ca","Type":"ContainerStarted","Data":"66e3fd7b3349ab823118640a7d1814f88e21cc4f4839427e764cab8c26290ee9"} Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.654216 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-logging/collector-ghzsk" podStartSLOduration=25.453212427 podStartE2EDuration="32.654190241s" podCreationTimestamp="2025-12-03 17:28:41 +0000 UTC" firstStartedPulling="2025-12-03 17:28:42.050325831 +0000 UTC m=+938.867796790" lastFinishedPulling="2025-12-03 17:28:49.251303615 +0000 UTC m=+946.068774604" observedRunningTime="2025-12-03 17:28:50.469630365 +0000 UTC m=+947.287101334" watchObservedRunningTime="2025-12-03 17:29:13.654190241 +0000 UTC m=+970.471661200" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.655238 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8"] Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.656393 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.658655 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.675087 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8"] Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.774489 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.774865 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf6f6\" (UniqueName: \"kubernetes.io/projected/7a1e8379-9aab-4e95-8c05-089edfed5d4b-kube-api-access-hf6f6\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.775166 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.877873 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.878493 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf6f6\" (UniqueName: \"kubernetes.io/projected/7a1e8379-9aab-4e95-8c05-089edfed5d4b-kube-api-access-hf6f6\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.878584 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.879735 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.880084 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.903640 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf6f6\" (UniqueName: \"kubernetes.io/projected/7a1e8379-9aab-4e95-8c05-089edfed5d4b-kube-api-access-hf6f6\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:13 crc kubenswrapper[4787]: I1203 17:29:13.975932 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:14 crc kubenswrapper[4787]: I1203 17:29:14.472140 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8"] Dec 03 17:29:14 crc kubenswrapper[4787]: I1203 17:29:14.622455 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" event={"ID":"7a1e8379-9aab-4e95-8c05-089edfed5d4b","Type":"ContainerStarted","Data":"740d5338762ded61eb1e3b17b1b22bc04dc2e0f51d97db540156f403460fb33f"} Dec 03 17:29:15 crc kubenswrapper[4787]: I1203 17:29:15.630995 4787 generic.go:334] "Generic (PLEG): container finished" podID="7a1e8379-9aab-4e95-8c05-089edfed5d4b" containerID="cbaebc566ea284a15eddd24651ea5bd0456fc467785ff0dffd93f74ec13759da" exitCode=0 Dec 03 17:29:15 crc kubenswrapper[4787]: I1203 17:29:15.631073 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" event={"ID":"7a1e8379-9aab-4e95-8c05-089edfed5d4b","Type":"ContainerDied","Data":"cbaebc566ea284a15eddd24651ea5bd0456fc467785ff0dffd93f74ec13759da"} Dec 03 17:29:17 crc kubenswrapper[4787]: I1203 17:29:17.646167 4787 generic.go:334] "Generic (PLEG): container finished" podID="7a1e8379-9aab-4e95-8c05-089edfed5d4b" containerID="a3a0a9a1a91629b679cea6d69e1ea7380959110560bd9e6c5676495fa99bffdc" exitCode=0 Dec 03 17:29:17 crc kubenswrapper[4787]: I1203 17:29:17.646226 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" event={"ID":"7a1e8379-9aab-4e95-8c05-089edfed5d4b","Type":"ContainerDied","Data":"a3a0a9a1a91629b679cea6d69e1ea7380959110560bd9e6c5676495fa99bffdc"} Dec 03 17:29:18 crc kubenswrapper[4787]: I1203 17:29:18.654490 4787 generic.go:334] "Generic (PLEG): container finished" podID="7a1e8379-9aab-4e95-8c05-089edfed5d4b" containerID="ca49aac09ae57119f34e8eb3617f83bd8983f99c9095c1d457fc7b9826da31a4" exitCode=0 Dec 03 17:29:18 crc kubenswrapper[4787]: I1203 17:29:18.654537 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" event={"ID":"7a1e8379-9aab-4e95-8c05-089edfed5d4b","Type":"ContainerDied","Data":"ca49aac09ae57119f34e8eb3617f83bd8983f99c9095c1d457fc7b9826da31a4"} Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.020726 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.188796 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-bundle\") pod \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.188875 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf6f6\" (UniqueName: \"kubernetes.io/projected/7a1e8379-9aab-4e95-8c05-089edfed5d4b-kube-api-access-hf6f6\") pod \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.188932 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-util\") pod \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\" (UID: \"7a1e8379-9aab-4e95-8c05-089edfed5d4b\") " Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.190190 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-bundle" (OuterVolumeSpecName: "bundle") pod "7a1e8379-9aab-4e95-8c05-089edfed5d4b" (UID: "7a1e8379-9aab-4e95-8c05-089edfed5d4b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.196786 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a1e8379-9aab-4e95-8c05-089edfed5d4b-kube-api-access-hf6f6" (OuterVolumeSpecName: "kube-api-access-hf6f6") pod "7a1e8379-9aab-4e95-8c05-089edfed5d4b" (UID: "7a1e8379-9aab-4e95-8c05-089edfed5d4b"). InnerVolumeSpecName "kube-api-access-hf6f6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.290874 4787 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.290916 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf6f6\" (UniqueName: \"kubernetes.io/projected/7a1e8379-9aab-4e95-8c05-089edfed5d4b-kube-api-access-hf6f6\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.417953 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-util" (OuterVolumeSpecName: "util") pod "7a1e8379-9aab-4e95-8c05-089edfed5d4b" (UID: "7a1e8379-9aab-4e95-8c05-089edfed5d4b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.494802 4787 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7a1e8379-9aab-4e95-8c05-089edfed5d4b-util\") on node \"crc\" DevicePath \"\"" Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.668960 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" event={"ID":"7a1e8379-9aab-4e95-8c05-089edfed5d4b","Type":"ContainerDied","Data":"740d5338762ded61eb1e3b17b1b22bc04dc2e0f51d97db540156f403460fb33f"} Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.669043 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="740d5338762ded61eb1e3b17b1b22bc04dc2e0f51d97db540156f403460fb33f" Dec 03 17:29:20 crc kubenswrapper[4787]: I1203 17:29:20.669136 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8" Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.703110 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-fhrgj"] Dec 03 17:29:25 crc kubenswrapper[4787]: E1203 17:29:25.703950 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a1e8379-9aab-4e95-8c05-089edfed5d4b" containerName="pull" Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.703965 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a1e8379-9aab-4e95-8c05-089edfed5d4b" containerName="pull" Dec 03 17:29:25 crc kubenswrapper[4787]: E1203 17:29:25.703995 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a1e8379-9aab-4e95-8c05-089edfed5d4b" containerName="extract" Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.704003 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a1e8379-9aab-4e95-8c05-089edfed5d4b" containerName="extract" Dec 03 17:29:25 crc kubenswrapper[4787]: E1203 17:29:25.704014 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a1e8379-9aab-4e95-8c05-089edfed5d4b" containerName="util" Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.704038 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a1e8379-9aab-4e95-8c05-089edfed5d4b" containerName="util" Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.704180 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a1e8379-9aab-4e95-8c05-089edfed5d4b" containerName="extract" Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.704765 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fhrgj" Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.707138 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.707263 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-wvrcb" Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.707263 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.722370 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-fhrgj"] Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.873087 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5lw4\" (UniqueName: \"kubernetes.io/projected/e4096b9f-51a5-4bb4-860e-0f689a2d9f07-kube-api-access-w5lw4\") pod \"nmstate-operator-5b5b58f5c8-fhrgj\" (UID: \"e4096b9f-51a5-4bb4-860e-0f689a2d9f07\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fhrgj" Dec 03 17:29:25 crc kubenswrapper[4787]: I1203 17:29:25.974267 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5lw4\" (UniqueName: \"kubernetes.io/projected/e4096b9f-51a5-4bb4-860e-0f689a2d9f07-kube-api-access-w5lw4\") pod \"nmstate-operator-5b5b58f5c8-fhrgj\" (UID: \"e4096b9f-51a5-4bb4-860e-0f689a2d9f07\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fhrgj" Dec 03 17:29:26 crc kubenswrapper[4787]: I1203 17:29:25.997438 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5lw4\" (UniqueName: \"kubernetes.io/projected/e4096b9f-51a5-4bb4-860e-0f689a2d9f07-kube-api-access-w5lw4\") pod \"nmstate-operator-5b5b58f5c8-fhrgj\" (UID: \"e4096b9f-51a5-4bb4-860e-0f689a2d9f07\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fhrgj" Dec 03 17:29:26 crc kubenswrapper[4787]: I1203 17:29:26.031892 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fhrgj" Dec 03 17:29:26 crc kubenswrapper[4787]: I1203 17:29:26.518783 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-fhrgj"] Dec 03 17:29:26 crc kubenswrapper[4787]: I1203 17:29:26.713965 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fhrgj" event={"ID":"e4096b9f-51a5-4bb4-860e-0f689a2d9f07","Type":"ContainerStarted","Data":"ead0fc2387db402e1110e4059e1161302c73e7c65a61ecb997ff8c3bc0e1a9f6"} Dec 03 17:29:29 crc kubenswrapper[4787]: I1203 17:29:29.736120 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fhrgj" event={"ID":"e4096b9f-51a5-4bb4-860e-0f689a2d9f07","Type":"ContainerStarted","Data":"bcfe259886e703741905fb054cd0cc8177592ac6c97d4f3af0924d8662cef876"} Dec 03 17:29:34 crc kubenswrapper[4787]: I1203 17:29:34.878275 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-fhrgj" podStartSLOduration=7.478085094 podStartE2EDuration="9.878253565s" podCreationTimestamp="2025-12-03 17:29:25 +0000 UTC" firstStartedPulling="2025-12-03 17:29:26.532038958 +0000 UTC m=+983.349509927" lastFinishedPulling="2025-12-03 17:29:28.932207449 +0000 UTC m=+985.749678398" observedRunningTime="2025-12-03 17:29:29.758272134 +0000 UTC m=+986.575743243" watchObservedRunningTime="2025-12-03 17:29:34.878253565 +0000 UTC m=+991.695724524" Dec 03 17:29:34 crc kubenswrapper[4787]: I1203 17:29:34.881120 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp"] Dec 03 17:29:34 crc kubenswrapper[4787]: I1203 17:29:34.882393 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp" Dec 03 17:29:34 crc kubenswrapper[4787]: I1203 17:29:34.888232 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-bmlxz" Dec 03 17:29:34 crc kubenswrapper[4787]: I1203 17:29:34.929583 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4"] Dec 03 17:29:34 crc kubenswrapper[4787]: I1203 17:29:34.930452 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" Dec 03 17:29:34 crc kubenswrapper[4787]: I1203 17:29:34.931895 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 03 17:29:34 crc kubenswrapper[4787]: I1203 17:29:34.934707 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp"] Dec 03 17:29:34 crc kubenswrapper[4787]: I1203 17:29:34.945353 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-njhbf"] Dec 03 17:29:34 crc kubenswrapper[4787]: I1203 17:29:34.946166 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:34 crc kubenswrapper[4787]: I1203 17:29:34.950193 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4"] Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.024815 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n47hv\" (UniqueName: \"kubernetes.io/projected/2915592e-3cf5-43c3-a142-d89b98274df2-kube-api-access-n47hv\") pod \"nmstate-webhook-5f6d4c5ccb-b4gp4\" (UID: \"2915592e-3cf5-43c3-a142-d89b98274df2\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.024925 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2915592e-3cf5-43c3-a142-d89b98274df2-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-b4gp4\" (UID: \"2915592e-3cf5-43c3-a142-d89b98274df2\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.024957 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rhhcp\" (UniqueName: \"kubernetes.io/projected/2a021e2e-e7b3-44ed-8bc2-3008461b97b3-kube-api-access-rhhcp\") pod \"nmstate-metrics-7f946cbc9-p9brp\" (UID: \"2a021e2e-e7b3-44ed-8bc2-3008461b97b3\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.045954 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj"] Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.048059 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.060431 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.060672 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.060932 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-tckqd" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.071762 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj"] Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.126235 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f8584654-d601-42cc-98e4-1abd6fdbe848-ovs-socket\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.126293 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f8584654-d601-42cc-98e4-1abd6fdbe848-dbus-socket\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.126522 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n47hv\" (UniqueName: \"kubernetes.io/projected/2915592e-3cf5-43c3-a142-d89b98274df2-kube-api-access-n47hv\") pod \"nmstate-webhook-5f6d4c5ccb-b4gp4\" (UID: \"2915592e-3cf5-43c3-a142-d89b98274df2\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.126562 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v526d\" (UniqueName: \"kubernetes.io/projected/f8584654-d601-42cc-98e4-1abd6fdbe848-kube-api-access-v526d\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.126593 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2915592e-3cf5-43c3-a142-d89b98274df2-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-b4gp4\" (UID: \"2915592e-3cf5-43c3-a142-d89b98274df2\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.126620 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rhhcp\" (UniqueName: \"kubernetes.io/projected/2a021e2e-e7b3-44ed-8bc2-3008461b97b3-kube-api-access-rhhcp\") pod \"nmstate-metrics-7f946cbc9-p9brp\" (UID: \"2a021e2e-e7b3-44ed-8bc2-3008461b97b3\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.126686 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f8584654-d601-42cc-98e4-1abd6fdbe848-nmstate-lock\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.133735 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2915592e-3cf5-43c3-a142-d89b98274df2-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-b4gp4\" (UID: \"2915592e-3cf5-43c3-a142-d89b98274df2\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.143114 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rhhcp\" (UniqueName: \"kubernetes.io/projected/2a021e2e-e7b3-44ed-8bc2-3008461b97b3-kube-api-access-rhhcp\") pod \"nmstate-metrics-7f946cbc9-p9brp\" (UID: \"2a021e2e-e7b3-44ed-8bc2-3008461b97b3\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.146367 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n47hv\" (UniqueName: \"kubernetes.io/projected/2915592e-3cf5-43c3-a142-d89b98274df2-kube-api-access-n47hv\") pod \"nmstate-webhook-5f6d4c5ccb-b4gp4\" (UID: \"2915592e-3cf5-43c3-a142-d89b98274df2\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.228380 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce1a6a45-6907-432d-a9c6-3d1f1b9fad52-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-j2lgj\" (UID: \"ce1a6a45-6907-432d-a9c6-3d1f1b9fad52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.228814 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v526d\" (UniqueName: \"kubernetes.io/projected/f8584654-d601-42cc-98e4-1abd6fdbe848-kube-api-access-v526d\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.228861 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ce1a6a45-6907-432d-a9c6-3d1f1b9fad52-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-j2lgj\" (UID: \"ce1a6a45-6907-432d-a9c6-3d1f1b9fad52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.228891 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9bbzq\" (UniqueName: \"kubernetes.io/projected/ce1a6a45-6907-432d-a9c6-3d1f1b9fad52-kube-api-access-9bbzq\") pod \"nmstate-console-plugin-7fbb5f6569-j2lgj\" (UID: \"ce1a6a45-6907-432d-a9c6-3d1f1b9fad52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.228918 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f8584654-d601-42cc-98e4-1abd6fdbe848-nmstate-lock\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.228953 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f8584654-d601-42cc-98e4-1abd6fdbe848-ovs-socket\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.228972 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f8584654-d601-42cc-98e4-1abd6fdbe848-dbus-socket\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.229319 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/f8584654-d601-42cc-98e4-1abd6fdbe848-dbus-socket\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.229524 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/f8584654-d601-42cc-98e4-1abd6fdbe848-nmstate-lock\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.229524 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/f8584654-d601-42cc-98e4-1abd6fdbe848-ovs-socket\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.241864 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-85474b5df5-nrlj4"] Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.242807 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.244977 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.252805 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v526d\" (UniqueName: \"kubernetes.io/projected/f8584654-d601-42cc-98e4-1abd6fdbe848-kube-api-access-v526d\") pod \"nmstate-handler-njhbf\" (UID: \"f8584654-d601-42cc-98e4-1abd6fdbe848\") " pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.257596 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.268727 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.274605 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-85474b5df5-nrlj4"] Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.329699 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-serving-cert\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.329751 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kthzp\" (UniqueName: \"kubernetes.io/projected/e4903b58-00fb-44e2-a2a1-60f89894b73c-kube-api-access-kthzp\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.329778 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-oauth-config\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.329795 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-oauth-serving-cert\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.329815 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-trusted-ca-bundle\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.329841 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce1a6a45-6907-432d-a9c6-3d1f1b9fad52-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-j2lgj\" (UID: \"ce1a6a45-6907-432d-a9c6-3d1f1b9fad52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.329859 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-config\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.329885 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-service-ca\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.329912 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ce1a6a45-6907-432d-a9c6-3d1f1b9fad52-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-j2lgj\" (UID: \"ce1a6a45-6907-432d-a9c6-3d1f1b9fad52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.329941 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9bbzq\" (UniqueName: \"kubernetes.io/projected/ce1a6a45-6907-432d-a9c6-3d1f1b9fad52-kube-api-access-9bbzq\") pod \"nmstate-console-plugin-7fbb5f6569-j2lgj\" (UID: \"ce1a6a45-6907-432d-a9c6-3d1f1b9fad52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.332356 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/ce1a6a45-6907-432d-a9c6-3d1f1b9fad52-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-j2lgj\" (UID: \"ce1a6a45-6907-432d-a9c6-3d1f1b9fad52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.336321 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/ce1a6a45-6907-432d-a9c6-3d1f1b9fad52-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-j2lgj\" (UID: \"ce1a6a45-6907-432d-a9c6-3d1f1b9fad52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.346768 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9bbzq\" (UniqueName: \"kubernetes.io/projected/ce1a6a45-6907-432d-a9c6-3d1f1b9fad52-kube-api-access-9bbzq\") pod \"nmstate-console-plugin-7fbb5f6569-j2lgj\" (UID: \"ce1a6a45-6907-432d-a9c6-3d1f1b9fad52\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.375279 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.431438 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-serving-cert\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.431521 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kthzp\" (UniqueName: \"kubernetes.io/projected/e4903b58-00fb-44e2-a2a1-60f89894b73c-kube-api-access-kthzp\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.431557 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-oauth-config\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.431580 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-oauth-serving-cert\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.431600 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-trusted-ca-bundle\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.431631 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-config\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.431665 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-service-ca\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.433103 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-oauth-serving-cert\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.434335 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-service-ca\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.435181 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-config\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.435406 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-trusted-ca-bundle\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.436175 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-oauth-config\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.436800 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-serving-cert\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.446461 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kthzp\" (UniqueName: \"kubernetes.io/projected/e4903b58-00fb-44e2-a2a1-60f89894b73c-kube-api-access-kthzp\") pod \"console-85474b5df5-nrlj4\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.612259 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj"] Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.655700 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.711322 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp"] Dec 03 17:29:35 crc kubenswrapper[4787]: W1203 17:29:35.730410 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a021e2e_e7b3_44ed_8bc2_3008461b97b3.slice/crio-0c4211591c7a19e2a4daebf8beb8f39bca885bebd80fe5fd53d04cb8c5adbfe8 WatchSource:0}: Error finding container 0c4211591c7a19e2a4daebf8beb8f39bca885bebd80fe5fd53d04cb8c5adbfe8: Status 404 returned error can't find the container with id 0c4211591c7a19e2a4daebf8beb8f39bca885bebd80fe5fd53d04cb8c5adbfe8 Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.755137 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4"] Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.786212 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-njhbf" event={"ID":"f8584654-d601-42cc-98e4-1abd6fdbe848","Type":"ContainerStarted","Data":"80b196a073de045420f5e9ecb14d59041096bbc7efd0116056130f438a484bb3"} Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.787954 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp" event={"ID":"2a021e2e-e7b3-44ed-8bc2-3008461b97b3","Type":"ContainerStarted","Data":"0c4211591c7a19e2a4daebf8beb8f39bca885bebd80fe5fd53d04cb8c5adbfe8"} Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.789926 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" event={"ID":"2915592e-3cf5-43c3-a142-d89b98274df2","Type":"ContainerStarted","Data":"9f82f9a3f49e0f16a7e42eb0d490ed3ad417fa2c4edbc6e8adb1bea265776b8c"} Dec 03 17:29:35 crc kubenswrapper[4787]: I1203 17:29:35.790996 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" event={"ID":"ce1a6a45-6907-432d-a9c6-3d1f1b9fad52","Type":"ContainerStarted","Data":"c37f794f71e850b9e056be21e01ae464bbe7681d2de263f263422de12787938b"} Dec 03 17:29:36 crc kubenswrapper[4787]: I1203 17:29:36.141615 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-85474b5df5-nrlj4"] Dec 03 17:29:36 crc kubenswrapper[4787]: W1203 17:29:36.148460 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode4903b58_00fb_44e2_a2a1_60f89894b73c.slice/crio-b75e6512fd9f5ac4f775f122247c3548c9f7604088aca906209ccd78ae9cbea0 WatchSource:0}: Error finding container b75e6512fd9f5ac4f775f122247c3548c9f7604088aca906209ccd78ae9cbea0: Status 404 returned error can't find the container with id b75e6512fd9f5ac4f775f122247c3548c9f7604088aca906209ccd78ae9cbea0 Dec 03 17:29:36 crc kubenswrapper[4787]: I1203 17:29:36.799278 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-85474b5df5-nrlj4" event={"ID":"e4903b58-00fb-44e2-a2a1-60f89894b73c","Type":"ContainerStarted","Data":"b75e6512fd9f5ac4f775f122247c3548c9f7604088aca906209ccd78ae9cbea0"} Dec 03 17:29:37 crc kubenswrapper[4787]: I1203 17:29:37.822115 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-85474b5df5-nrlj4" event={"ID":"e4903b58-00fb-44e2-a2a1-60f89894b73c","Type":"ContainerStarted","Data":"9c805ac3f15f3ac1131c700a4c872396df24d2d62ff0e5da005e0ec0c8e6132b"} Dec 03 17:29:39 crc kubenswrapper[4787]: I1203 17:29:39.837242 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" event={"ID":"2915592e-3cf5-43c3-a142-d89b98274df2","Type":"ContainerStarted","Data":"12a62c151841828ac1d9503e40189666c3036e125df983a54838f04eca4184e5"} Dec 03 17:29:39 crc kubenswrapper[4787]: I1203 17:29:39.837909 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" Dec 03 17:29:39 crc kubenswrapper[4787]: I1203 17:29:39.840610 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" event={"ID":"ce1a6a45-6907-432d-a9c6-3d1f1b9fad52","Type":"ContainerStarted","Data":"bf639f47f146bd3f07a39ad8074303e132c65567eda13f7b161463e967b8d12d"} Dec 03 17:29:39 crc kubenswrapper[4787]: I1203 17:29:39.842410 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-njhbf" event={"ID":"f8584654-d601-42cc-98e4-1abd6fdbe848","Type":"ContainerStarted","Data":"f4b8c9eb3a208c95621b497f34b2577bfdbd85d644d1a2e1695f9b295930cdd1"} Dec 03 17:29:39 crc kubenswrapper[4787]: I1203 17:29:39.842550 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:39 crc kubenswrapper[4787]: I1203 17:29:39.843937 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp" event={"ID":"2a021e2e-e7b3-44ed-8bc2-3008461b97b3","Type":"ContainerStarted","Data":"be58a74bac377caa5c5a2b567f552adc3634ac62625bd3fce9829b833fc1a3c0"} Dec 03 17:29:39 crc kubenswrapper[4787]: I1203 17:29:39.859697 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-85474b5df5-nrlj4" podStartSLOduration=4.859676817 podStartE2EDuration="4.859676817s" podCreationTimestamp="2025-12-03 17:29:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:29:37.842635493 +0000 UTC m=+994.660106562" watchObservedRunningTime="2025-12-03 17:29:39.859676817 +0000 UTC m=+996.677147766" Dec 03 17:29:39 crc kubenswrapper[4787]: I1203 17:29:39.863787 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" podStartSLOduration=2.435335978 podStartE2EDuration="5.863766546s" podCreationTimestamp="2025-12-03 17:29:34 +0000 UTC" firstStartedPulling="2025-12-03 17:29:35.780005826 +0000 UTC m=+992.597476785" lastFinishedPulling="2025-12-03 17:29:39.208436394 +0000 UTC m=+996.025907353" observedRunningTime="2025-12-03 17:29:39.858298761 +0000 UTC m=+996.675769740" watchObservedRunningTime="2025-12-03 17:29:39.863766546 +0000 UTC m=+996.681237495" Dec 03 17:29:39 crc kubenswrapper[4787]: I1203 17:29:39.895699 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-j2lgj" podStartSLOduration=1.792746561 podStartE2EDuration="4.895675476s" podCreationTimestamp="2025-12-03 17:29:35 +0000 UTC" firstStartedPulling="2025-12-03 17:29:35.621835456 +0000 UTC m=+992.439306415" lastFinishedPulling="2025-12-03 17:29:38.724764371 +0000 UTC m=+995.542235330" observedRunningTime="2025-12-03 17:29:39.889423059 +0000 UTC m=+996.706894038" watchObservedRunningTime="2025-12-03 17:29:39.895675476 +0000 UTC m=+996.713146435" Dec 03 17:29:41 crc kubenswrapper[4787]: I1203 17:29:41.857717 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp" event={"ID":"2a021e2e-e7b3-44ed-8bc2-3008461b97b3","Type":"ContainerStarted","Data":"b15b82de89fde2826ae73fc7fb6921fe849cac4205073b6ca3446816612dc200"} Dec 03 17:29:41 crc kubenswrapper[4787]: I1203 17:29:41.873576 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-njhbf" podStartSLOduration=4.476432442 podStartE2EDuration="7.873556667s" podCreationTimestamp="2025-12-03 17:29:34 +0000 UTC" firstStartedPulling="2025-12-03 17:29:35.326693161 +0000 UTC m=+992.144164120" lastFinishedPulling="2025-12-03 17:29:38.723817386 +0000 UTC m=+995.541288345" observedRunningTime="2025-12-03 17:29:39.912556425 +0000 UTC m=+996.730027384" watchObservedRunningTime="2025-12-03 17:29:41.873556667 +0000 UTC m=+998.691027626" Dec 03 17:29:41 crc kubenswrapper[4787]: I1203 17:29:41.876514 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-p9brp" podStartSLOduration=2.47977397 podStartE2EDuration="7.876503635s" podCreationTimestamp="2025-12-03 17:29:34 +0000 UTC" firstStartedPulling="2025-12-03 17:29:35.732406579 +0000 UTC m=+992.549877538" lastFinishedPulling="2025-12-03 17:29:41.129136244 +0000 UTC m=+997.946607203" observedRunningTime="2025-12-03 17:29:41.872821547 +0000 UTC m=+998.690292506" watchObservedRunningTime="2025-12-03 17:29:41.876503635 +0000 UTC m=+998.693974594" Dec 03 17:29:45 crc kubenswrapper[4787]: I1203 17:29:45.306984 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-njhbf" Dec 03 17:29:45 crc kubenswrapper[4787]: I1203 17:29:45.655779 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:45 crc kubenswrapper[4787]: I1203 17:29:45.656164 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:45 crc kubenswrapper[4787]: I1203 17:29:45.661849 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:45 crc kubenswrapper[4787]: I1203 17:29:45.890836 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:29:45 crc kubenswrapper[4787]: I1203 17:29:45.970599 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-nm7m8"] Dec 03 17:29:48 crc kubenswrapper[4787]: I1203 17:29:48.989566 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:29:48 crc kubenswrapper[4787]: I1203 17:29:48.990607 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:29:55 crc kubenswrapper[4787]: I1203 17:29:55.263841 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-b4gp4" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.153252 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d"] Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.154844 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.156866 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.157041 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.162944 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d"] Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.270059 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n97jr\" (UniqueName: \"kubernetes.io/projected/a5b1845b-dcdb-4fb2-983a-4018aa964656-kube-api-access-n97jr\") pod \"collect-profiles-29413050-jgj7d\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.270185 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5b1845b-dcdb-4fb2-983a-4018aa964656-config-volume\") pod \"collect-profiles-29413050-jgj7d\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.270222 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5b1845b-dcdb-4fb2-983a-4018aa964656-secret-volume\") pod \"collect-profiles-29413050-jgj7d\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.371092 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n97jr\" (UniqueName: \"kubernetes.io/projected/a5b1845b-dcdb-4fb2-983a-4018aa964656-kube-api-access-n97jr\") pod \"collect-profiles-29413050-jgj7d\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.371146 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5b1845b-dcdb-4fb2-983a-4018aa964656-config-volume\") pod \"collect-profiles-29413050-jgj7d\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.371171 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5b1845b-dcdb-4fb2-983a-4018aa964656-secret-volume\") pod \"collect-profiles-29413050-jgj7d\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.373180 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5b1845b-dcdb-4fb2-983a-4018aa964656-config-volume\") pod \"collect-profiles-29413050-jgj7d\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.384286 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5b1845b-dcdb-4fb2-983a-4018aa964656-secret-volume\") pod \"collect-profiles-29413050-jgj7d\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.389594 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n97jr\" (UniqueName: \"kubernetes.io/projected/a5b1845b-dcdb-4fb2-983a-4018aa964656-kube-api-access-n97jr\") pod \"collect-profiles-29413050-jgj7d\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.472745 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.918425 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d"] Dec 03 17:30:00 crc kubenswrapper[4787]: W1203 17:30:00.924636 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice/crio-f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7 WatchSource:0}: Error finding container f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7: Status 404 returned error can't find the container with id f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7 Dec 03 17:30:00 crc kubenswrapper[4787]: I1203 17:30:00.991096 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" event={"ID":"a5b1845b-dcdb-4fb2-983a-4018aa964656","Type":"ContainerStarted","Data":"f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7"} Dec 03 17:30:02 crc kubenswrapper[4787]: I1203 17:30:02.003052 4787 generic.go:334] "Generic (PLEG): container finished" podID="a5b1845b-dcdb-4fb2-983a-4018aa964656" containerID="017e800dc669ff109cf880c14e980f50c5d5c95a101d8fb12f9a6a7631d30b85" exitCode=0 Dec 03 17:30:02 crc kubenswrapper[4787]: I1203 17:30:02.003193 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" event={"ID":"a5b1845b-dcdb-4fb2-983a-4018aa964656","Type":"ContainerDied","Data":"017e800dc669ff109cf880c14e980f50c5d5c95a101d8fb12f9a6a7631d30b85"} Dec 03 17:30:03 crc kubenswrapper[4787]: I1203 17:30:03.355654 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:03 crc kubenswrapper[4787]: I1203 17:30:03.517645 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5b1845b-dcdb-4fb2-983a-4018aa964656-secret-volume\") pod \"a5b1845b-dcdb-4fb2-983a-4018aa964656\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " Dec 03 17:30:03 crc kubenswrapper[4787]: I1203 17:30:03.517903 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5b1845b-dcdb-4fb2-983a-4018aa964656-config-volume\") pod \"a5b1845b-dcdb-4fb2-983a-4018aa964656\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " Dec 03 17:30:03 crc kubenswrapper[4787]: I1203 17:30:03.517957 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n97jr\" (UniqueName: \"kubernetes.io/projected/a5b1845b-dcdb-4fb2-983a-4018aa964656-kube-api-access-n97jr\") pod \"a5b1845b-dcdb-4fb2-983a-4018aa964656\" (UID: \"a5b1845b-dcdb-4fb2-983a-4018aa964656\") " Dec 03 17:30:03 crc kubenswrapper[4787]: I1203 17:30:03.520264 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a5b1845b-dcdb-4fb2-983a-4018aa964656-config-volume" (OuterVolumeSpecName: "config-volume") pod "a5b1845b-dcdb-4fb2-983a-4018aa964656" (UID: "a5b1845b-dcdb-4fb2-983a-4018aa964656"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:30:03 crc kubenswrapper[4787]: I1203 17:30:03.524988 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5b1845b-dcdb-4fb2-983a-4018aa964656-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a5b1845b-dcdb-4fb2-983a-4018aa964656" (UID: "a5b1845b-dcdb-4fb2-983a-4018aa964656"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:30:03 crc kubenswrapper[4787]: I1203 17:30:03.525454 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5b1845b-dcdb-4fb2-983a-4018aa964656-kube-api-access-n97jr" (OuterVolumeSpecName: "kube-api-access-n97jr") pod "a5b1845b-dcdb-4fb2-983a-4018aa964656" (UID: "a5b1845b-dcdb-4fb2-983a-4018aa964656"). InnerVolumeSpecName "kube-api-access-n97jr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:30:03 crc kubenswrapper[4787]: I1203 17:30:03.620947 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a5b1845b-dcdb-4fb2-983a-4018aa964656-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:03 crc kubenswrapper[4787]: I1203 17:30:03.621037 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a5b1845b-dcdb-4fb2-983a-4018aa964656-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:03 crc kubenswrapper[4787]: I1203 17:30:03.621057 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n97jr\" (UniqueName: \"kubernetes.io/projected/a5b1845b-dcdb-4fb2-983a-4018aa964656-kube-api-access-n97jr\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:04 crc kubenswrapper[4787]: I1203 17:30:04.026286 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" event={"ID":"a5b1845b-dcdb-4fb2-983a-4018aa964656","Type":"ContainerDied","Data":"f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7"} Dec 03 17:30:04 crc kubenswrapper[4787]: I1203 17:30:04.027271 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7" Dec 03 17:30:04 crc kubenswrapper[4787]: I1203 17:30:04.026380 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d" Dec 03 17:30:07 crc kubenswrapper[4787]: E1203 17:30:07.473913 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice/crio-f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice\": RecentStats: unable to find data in memory cache]" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.536870 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8"] Dec 03 17:30:10 crc kubenswrapper[4787]: E1203 17:30:10.539544 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5b1845b-dcdb-4fb2-983a-4018aa964656" containerName="collect-profiles" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.539642 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5b1845b-dcdb-4fb2-983a-4018aa964656" containerName="collect-profiles" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.539825 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5b1845b-dcdb-4fb2-983a-4018aa964656" containerName="collect-profiles" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.541063 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.548528 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.560269 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8"] Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.647933 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjdd9\" (UniqueName: \"kubernetes.io/projected/7bed6e06-e6f3-4611-90fd-8963706e05d5-kube-api-access-sjdd9\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.648284 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.648451 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.750038 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.750367 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.750551 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjdd9\" (UniqueName: \"kubernetes.io/projected/7bed6e06-e6f3-4611-90fd-8963706e05d5-kube-api-access-sjdd9\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.750734 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.751190 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.777222 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjdd9\" (UniqueName: \"kubernetes.io/projected/7bed6e06-e6f3-4611-90fd-8963706e05d5-kube-api-access-sjdd9\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:10 crc kubenswrapper[4787]: I1203 17:30:10.861042 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.036415 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-nm7m8" podUID="a66535aa-a358-4570-ae9f-5dac58121ef2" containerName="console" containerID="cri-o://7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934" gracePeriod=15 Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.353511 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8"] Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.469099 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-nm7m8_a66535aa-a358-4570-ae9f-5dac58121ef2/console/0.log" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.469592 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.569257 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-oauth-serving-cert\") pod \"a66535aa-a358-4570-ae9f-5dac58121ef2\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.569430 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-trusted-ca-bundle\") pod \"a66535aa-a358-4570-ae9f-5dac58121ef2\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.569492 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-serving-cert\") pod \"a66535aa-a358-4570-ae9f-5dac58121ef2\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.569516 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-service-ca\") pod \"a66535aa-a358-4570-ae9f-5dac58121ef2\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.569533 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-console-config\") pod \"a66535aa-a358-4570-ae9f-5dac58121ef2\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.569562 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzgfv\" (UniqueName: \"kubernetes.io/projected/a66535aa-a358-4570-ae9f-5dac58121ef2-kube-api-access-bzgfv\") pod \"a66535aa-a358-4570-ae9f-5dac58121ef2\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.569590 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-oauth-config\") pod \"a66535aa-a358-4570-ae9f-5dac58121ef2\" (UID: \"a66535aa-a358-4570-ae9f-5dac58121ef2\") " Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.581564 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-console-config" (OuterVolumeSpecName: "console-config") pod "a66535aa-a358-4570-ae9f-5dac58121ef2" (UID: "a66535aa-a358-4570-ae9f-5dac58121ef2"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.582648 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "a66535aa-a358-4570-ae9f-5dac58121ef2" (UID: "a66535aa-a358-4570-ae9f-5dac58121ef2"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.583204 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-service-ca" (OuterVolumeSpecName: "service-ca") pod "a66535aa-a358-4570-ae9f-5dac58121ef2" (UID: "a66535aa-a358-4570-ae9f-5dac58121ef2"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.583245 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "a66535aa-a358-4570-ae9f-5dac58121ef2" (UID: "a66535aa-a358-4570-ae9f-5dac58121ef2"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.583743 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "a66535aa-a358-4570-ae9f-5dac58121ef2" (UID: "a66535aa-a358-4570-ae9f-5dac58121ef2"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.585540 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a66535aa-a358-4570-ae9f-5dac58121ef2-kube-api-access-bzgfv" (OuterVolumeSpecName: "kube-api-access-bzgfv") pod "a66535aa-a358-4570-ae9f-5dac58121ef2" (UID: "a66535aa-a358-4570-ae9f-5dac58121ef2"). InnerVolumeSpecName "kube-api-access-bzgfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.585888 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "a66535aa-a358-4570-ae9f-5dac58121ef2" (UID: "a66535aa-a358-4570-ae9f-5dac58121ef2"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.671589 4787 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.671624 4787 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.671633 4787 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.671643 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzgfv\" (UniqueName: \"kubernetes.io/projected/a66535aa-a358-4570-ae9f-5dac58121ef2-kube-api-access-bzgfv\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.671654 4787 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/a66535aa-a358-4570-ae9f-5dac58121ef2-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.671662 4787 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:11 crc kubenswrapper[4787]: I1203 17:30:11.671670 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a66535aa-a358-4570-ae9f-5dac58121ef2-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.091194 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-nm7m8_a66535aa-a358-4570-ae9f-5dac58121ef2/console/0.log" Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.091569 4787 generic.go:334] "Generic (PLEG): container finished" podID="a66535aa-a358-4570-ae9f-5dac58121ef2" containerID="7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934" exitCode=2 Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.091627 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nm7m8" event={"ID":"a66535aa-a358-4570-ae9f-5dac58121ef2","Type":"ContainerDied","Data":"7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934"} Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.091656 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-nm7m8" event={"ID":"a66535aa-a358-4570-ae9f-5dac58121ef2","Type":"ContainerDied","Data":"ec9e06a0eb112a7de167d3880784823d62b3a878cc9564bc48ece361534c05f6"} Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.091671 4787 scope.go:117] "RemoveContainer" containerID="7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934" Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.091674 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-nm7m8" Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.093849 4787 generic.go:334] "Generic (PLEG): container finished" podID="7bed6e06-e6f3-4611-90fd-8963706e05d5" containerID="abc2953d03b9fbe99e5c3a8b5b64812044172896e8bf195f2effa576e3bb3e97" exitCode=0 Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.093913 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" event={"ID":"7bed6e06-e6f3-4611-90fd-8963706e05d5","Type":"ContainerDied","Data":"abc2953d03b9fbe99e5c3a8b5b64812044172896e8bf195f2effa576e3bb3e97"} Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.093953 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" event={"ID":"7bed6e06-e6f3-4611-90fd-8963706e05d5","Type":"ContainerStarted","Data":"33dcd7da727236e2c38f4d0b6dff9bf6a93bf145b9dae8392365629c3baa4015"} Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.096102 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.124535 4787 scope.go:117] "RemoveContainer" containerID="7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934" Dec 03 17:30:12 crc kubenswrapper[4787]: E1203 17:30:12.126218 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934\": container with ID starting with 7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934 not found: ID does not exist" containerID="7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934" Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.126276 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934"} err="failed to get container status \"7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934\": rpc error: code = NotFound desc = could not find container \"7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934\": container with ID starting with 7c5dc2a07101cd0bbefe8e3d466451c8fd83373c00128251c792de75b8f19934 not found: ID does not exist" Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.146270 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-nm7m8"] Dec 03 17:30:12 crc kubenswrapper[4787]: I1203 17:30:12.150182 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-nm7m8"] Dec 03 17:30:13 crc kubenswrapper[4787]: I1203 17:30:13.780496 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a66535aa-a358-4570-ae9f-5dac58121ef2" path="/var/lib/kubelet/pods/a66535aa-a358-4570-ae9f-5dac58121ef2/volumes" Dec 03 17:30:14 crc kubenswrapper[4787]: I1203 17:30:14.113167 4787 generic.go:334] "Generic (PLEG): container finished" podID="7bed6e06-e6f3-4611-90fd-8963706e05d5" containerID="72d3c0ff411d75dc752191ae86cd287990ceb980c9aa664d08ef6df8305e39c4" exitCode=0 Dec 03 17:30:14 crc kubenswrapper[4787]: I1203 17:30:14.113217 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" event={"ID":"7bed6e06-e6f3-4611-90fd-8963706e05d5","Type":"ContainerDied","Data":"72d3c0ff411d75dc752191ae86cd287990ceb980c9aa664d08ef6df8305e39c4"} Dec 03 17:30:15 crc kubenswrapper[4787]: I1203 17:30:15.122471 4787 generic.go:334] "Generic (PLEG): container finished" podID="7bed6e06-e6f3-4611-90fd-8963706e05d5" containerID="e3502a31757d39250334bc22d37ba815641a0d4cc5b62ef37c67a9a13861c64b" exitCode=0 Dec 03 17:30:15 crc kubenswrapper[4787]: I1203 17:30:15.122549 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" event={"ID":"7bed6e06-e6f3-4611-90fd-8963706e05d5","Type":"ContainerDied","Data":"e3502a31757d39250334bc22d37ba815641a0d4cc5b62ef37c67a9a13861c64b"} Dec 03 17:30:16 crc kubenswrapper[4787]: I1203 17:30:16.516377 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:16 crc kubenswrapper[4787]: I1203 17:30:16.643892 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-util\") pod \"7bed6e06-e6f3-4611-90fd-8963706e05d5\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " Dec 03 17:30:16 crc kubenswrapper[4787]: I1203 17:30:16.643952 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-bundle\") pod \"7bed6e06-e6f3-4611-90fd-8963706e05d5\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " Dec 03 17:30:16 crc kubenswrapper[4787]: I1203 17:30:16.644098 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjdd9\" (UniqueName: \"kubernetes.io/projected/7bed6e06-e6f3-4611-90fd-8963706e05d5-kube-api-access-sjdd9\") pod \"7bed6e06-e6f3-4611-90fd-8963706e05d5\" (UID: \"7bed6e06-e6f3-4611-90fd-8963706e05d5\") " Dec 03 17:30:16 crc kubenswrapper[4787]: I1203 17:30:16.645161 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-bundle" (OuterVolumeSpecName: "bundle") pod "7bed6e06-e6f3-4611-90fd-8963706e05d5" (UID: "7bed6e06-e6f3-4611-90fd-8963706e05d5"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:30:16 crc kubenswrapper[4787]: I1203 17:30:16.651228 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bed6e06-e6f3-4611-90fd-8963706e05d5-kube-api-access-sjdd9" (OuterVolumeSpecName: "kube-api-access-sjdd9") pod "7bed6e06-e6f3-4611-90fd-8963706e05d5" (UID: "7bed6e06-e6f3-4611-90fd-8963706e05d5"). InnerVolumeSpecName "kube-api-access-sjdd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:30:16 crc kubenswrapper[4787]: I1203 17:30:16.658295 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-util" (OuterVolumeSpecName: "util") pod "7bed6e06-e6f3-4611-90fd-8963706e05d5" (UID: "7bed6e06-e6f3-4611-90fd-8963706e05d5"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:30:16 crc kubenswrapper[4787]: I1203 17:30:16.745656 4787 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:16 crc kubenswrapper[4787]: I1203 17:30:16.745694 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjdd9\" (UniqueName: \"kubernetes.io/projected/7bed6e06-e6f3-4611-90fd-8963706e05d5-kube-api-access-sjdd9\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:16 crc kubenswrapper[4787]: I1203 17:30:16.745705 4787 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/7bed6e06-e6f3-4611-90fd-8963706e05d5-util\") on node \"crc\" DevicePath \"\"" Dec 03 17:30:17 crc kubenswrapper[4787]: I1203 17:30:17.139448 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" event={"ID":"7bed6e06-e6f3-4611-90fd-8963706e05d5","Type":"ContainerDied","Data":"33dcd7da727236e2c38f4d0b6dff9bf6a93bf145b9dae8392365629c3baa4015"} Dec 03 17:30:17 crc kubenswrapper[4787]: I1203 17:30:17.139523 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="33dcd7da727236e2c38f4d0b6dff9bf6a93bf145b9dae8392365629c3baa4015" Dec 03 17:30:17 crc kubenswrapper[4787]: I1203 17:30:17.139543 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8" Dec 03 17:30:17 crc kubenswrapper[4787]: E1203 17:30:17.621171 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice/crio-f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7\": RecentStats: unable to find data in memory cache]" Dec 03 17:30:18 crc kubenswrapper[4787]: I1203 17:30:18.990912 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:30:18 crc kubenswrapper[4787]: I1203 17:30:18.991284 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.618081 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb"] Dec 03 17:30:25 crc kubenswrapper[4787]: E1203 17:30:25.619085 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bed6e06-e6f3-4611-90fd-8963706e05d5" containerName="util" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.619101 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bed6e06-e6f3-4611-90fd-8963706e05d5" containerName="util" Dec 03 17:30:25 crc kubenswrapper[4787]: E1203 17:30:25.619114 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bed6e06-e6f3-4611-90fd-8963706e05d5" containerName="pull" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.619122 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bed6e06-e6f3-4611-90fd-8963706e05d5" containerName="pull" Dec 03 17:30:25 crc kubenswrapper[4787]: E1203 17:30:25.619132 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a66535aa-a358-4570-ae9f-5dac58121ef2" containerName="console" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.619140 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a66535aa-a358-4570-ae9f-5dac58121ef2" containerName="console" Dec 03 17:30:25 crc kubenswrapper[4787]: E1203 17:30:25.619164 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7bed6e06-e6f3-4611-90fd-8963706e05d5" containerName="extract" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.619172 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7bed6e06-e6f3-4611-90fd-8963706e05d5" containerName="extract" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.619321 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a66535aa-a358-4570-ae9f-5dac58121ef2" containerName="console" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.619332 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7bed6e06-e6f3-4611-90fd-8963706e05d5" containerName="extract" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.619919 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.636414 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.636670 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.636832 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.636921 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.636986 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-tpg5h" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.651776 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb"] Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.786722 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe122b60-4167-40b6-8562-1b52112f44f8-webhook-cert\") pod \"metallb-operator-controller-manager-9b64dcbdc-qzmjb\" (UID: \"fe122b60-4167-40b6-8562-1b52112f44f8\") " pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.786814 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbxnp\" (UniqueName: \"kubernetes.io/projected/fe122b60-4167-40b6-8562-1b52112f44f8-kube-api-access-tbxnp\") pod \"metallb-operator-controller-manager-9b64dcbdc-qzmjb\" (UID: \"fe122b60-4167-40b6-8562-1b52112f44f8\") " pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.787004 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe122b60-4167-40b6-8562-1b52112f44f8-apiservice-cert\") pod \"metallb-operator-controller-manager-9b64dcbdc-qzmjb\" (UID: \"fe122b60-4167-40b6-8562-1b52112f44f8\") " pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.888168 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe122b60-4167-40b6-8562-1b52112f44f8-apiservice-cert\") pod \"metallb-operator-controller-manager-9b64dcbdc-qzmjb\" (UID: \"fe122b60-4167-40b6-8562-1b52112f44f8\") " pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.888267 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe122b60-4167-40b6-8562-1b52112f44f8-webhook-cert\") pod \"metallb-operator-controller-manager-9b64dcbdc-qzmjb\" (UID: \"fe122b60-4167-40b6-8562-1b52112f44f8\") " pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.888301 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbxnp\" (UniqueName: \"kubernetes.io/projected/fe122b60-4167-40b6-8562-1b52112f44f8-kube-api-access-tbxnp\") pod \"metallb-operator-controller-manager-9b64dcbdc-qzmjb\" (UID: \"fe122b60-4167-40b6-8562-1b52112f44f8\") " pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.895158 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/fe122b60-4167-40b6-8562-1b52112f44f8-apiservice-cert\") pod \"metallb-operator-controller-manager-9b64dcbdc-qzmjb\" (UID: \"fe122b60-4167-40b6-8562-1b52112f44f8\") " pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.895854 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/fe122b60-4167-40b6-8562-1b52112f44f8-webhook-cert\") pod \"metallb-operator-controller-manager-9b64dcbdc-qzmjb\" (UID: \"fe122b60-4167-40b6-8562-1b52112f44f8\") " pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.930919 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbxnp\" (UniqueName: \"kubernetes.io/projected/fe122b60-4167-40b6-8562-1b52112f44f8-kube-api-access-tbxnp\") pod \"metallb-operator-controller-manager-9b64dcbdc-qzmjb\" (UID: \"fe122b60-4167-40b6-8562-1b52112f44f8\") " pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:25 crc kubenswrapper[4787]: I1203 17:30:25.936407 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.044640 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5"] Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.045504 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.054389 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.054644 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-hl87z" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.054775 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.095812 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5"] Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.197581 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/200c13aa-bae6-4b29-b6b1-35b625ce2dfc-apiservice-cert\") pod \"metallb-operator-webhook-server-86b8bcb77d-fd8v5\" (UID: \"200c13aa-bae6-4b29-b6b1-35b625ce2dfc\") " pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.197644 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/200c13aa-bae6-4b29-b6b1-35b625ce2dfc-webhook-cert\") pod \"metallb-operator-webhook-server-86b8bcb77d-fd8v5\" (UID: \"200c13aa-bae6-4b29-b6b1-35b625ce2dfc\") " pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.197731 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnzsj\" (UniqueName: \"kubernetes.io/projected/200c13aa-bae6-4b29-b6b1-35b625ce2dfc-kube-api-access-lnzsj\") pod \"metallb-operator-webhook-server-86b8bcb77d-fd8v5\" (UID: \"200c13aa-bae6-4b29-b6b1-35b625ce2dfc\") " pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.299160 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/200c13aa-bae6-4b29-b6b1-35b625ce2dfc-apiservice-cert\") pod \"metallb-operator-webhook-server-86b8bcb77d-fd8v5\" (UID: \"200c13aa-bae6-4b29-b6b1-35b625ce2dfc\") " pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.299579 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/200c13aa-bae6-4b29-b6b1-35b625ce2dfc-webhook-cert\") pod \"metallb-operator-webhook-server-86b8bcb77d-fd8v5\" (UID: \"200c13aa-bae6-4b29-b6b1-35b625ce2dfc\") " pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.299661 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnzsj\" (UniqueName: \"kubernetes.io/projected/200c13aa-bae6-4b29-b6b1-35b625ce2dfc-kube-api-access-lnzsj\") pod \"metallb-operator-webhook-server-86b8bcb77d-fd8v5\" (UID: \"200c13aa-bae6-4b29-b6b1-35b625ce2dfc\") " pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.308518 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/200c13aa-bae6-4b29-b6b1-35b625ce2dfc-webhook-cert\") pod \"metallb-operator-webhook-server-86b8bcb77d-fd8v5\" (UID: \"200c13aa-bae6-4b29-b6b1-35b625ce2dfc\") " pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.316877 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnzsj\" (UniqueName: \"kubernetes.io/projected/200c13aa-bae6-4b29-b6b1-35b625ce2dfc-kube-api-access-lnzsj\") pod \"metallb-operator-webhook-server-86b8bcb77d-fd8v5\" (UID: \"200c13aa-bae6-4b29-b6b1-35b625ce2dfc\") " pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.324753 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/200c13aa-bae6-4b29-b6b1-35b625ce2dfc-apiservice-cert\") pod \"metallb-operator-webhook-server-86b8bcb77d-fd8v5\" (UID: \"200c13aa-bae6-4b29-b6b1-35b625ce2dfc\") " pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.373547 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.663324 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb"] Dec 03 17:30:26 crc kubenswrapper[4787]: I1203 17:30:26.870141 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5"] Dec 03 17:30:26 crc kubenswrapper[4787]: W1203 17:30:26.874206 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod200c13aa_bae6_4b29_b6b1_35b625ce2dfc.slice/crio-b63e1c21a80bc9227aaa73a1a18860c4c695fe0a8d26ebefc86dc79e20e7c887 WatchSource:0}: Error finding container b63e1c21a80bc9227aaa73a1a18860c4c695fe0a8d26ebefc86dc79e20e7c887: Status 404 returned error can't find the container with id b63e1c21a80bc9227aaa73a1a18860c4c695fe0a8d26ebefc86dc79e20e7c887 Dec 03 17:30:27 crc kubenswrapper[4787]: I1203 17:30:27.206393 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" event={"ID":"200c13aa-bae6-4b29-b6b1-35b625ce2dfc","Type":"ContainerStarted","Data":"b63e1c21a80bc9227aaa73a1a18860c4c695fe0a8d26ebefc86dc79e20e7c887"} Dec 03 17:30:27 crc kubenswrapper[4787]: I1203 17:30:27.207674 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" event={"ID":"fe122b60-4167-40b6-8562-1b52112f44f8","Type":"ContainerStarted","Data":"6cac62ffafcc42f199c9d05589054092980606bc6b7c7ffd893a42ececc9a377"} Dec 03 17:30:27 crc kubenswrapper[4787]: E1203 17:30:27.834189 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice/crio-f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice\": RecentStats: unable to find data in memory cache]" Dec 03 17:30:30 crc kubenswrapper[4787]: I1203 17:30:30.227383 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" event={"ID":"fe122b60-4167-40b6-8562-1b52112f44f8","Type":"ContainerStarted","Data":"d47b4f69dc44a842b728430c31095b9a7be4a931b16bb93818304110aecc6de7"} Dec 03 17:30:30 crc kubenswrapper[4787]: I1203 17:30:30.228293 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:30:30 crc kubenswrapper[4787]: I1203 17:30:30.257620 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" podStartSLOduration=1.913054596 podStartE2EDuration="5.257602251s" podCreationTimestamp="2025-12-03 17:30:25 +0000 UTC" firstStartedPulling="2025-12-03 17:30:26.676178775 +0000 UTC m=+1043.493649734" lastFinishedPulling="2025-12-03 17:30:30.02072643 +0000 UTC m=+1046.838197389" observedRunningTime="2025-12-03 17:30:30.246203379 +0000 UTC m=+1047.063674348" watchObservedRunningTime="2025-12-03 17:30:30.257602251 +0000 UTC m=+1047.075073210" Dec 03 17:30:32 crc kubenswrapper[4787]: I1203 17:30:32.240578 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" event={"ID":"200c13aa-bae6-4b29-b6b1-35b625ce2dfc","Type":"ContainerStarted","Data":"01ba4fbe7aa87000ec9ae32f24859cebe0da67994214e9236bd5968d4cb19678"} Dec 03 17:30:32 crc kubenswrapper[4787]: I1203 17:30:32.241230 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:32 crc kubenswrapper[4787]: I1203 17:30:32.257376 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" podStartSLOduration=1.111668863 podStartE2EDuration="6.257350323s" podCreationTimestamp="2025-12-03 17:30:26 +0000 UTC" firstStartedPulling="2025-12-03 17:30:26.87663468 +0000 UTC m=+1043.694105639" lastFinishedPulling="2025-12-03 17:30:32.02231614 +0000 UTC m=+1048.839787099" observedRunningTime="2025-12-03 17:30:32.254792255 +0000 UTC m=+1049.072263234" watchObservedRunningTime="2025-12-03 17:30:32.257350323 +0000 UTC m=+1049.074821292" Dec 03 17:30:38 crc kubenswrapper[4787]: E1203 17:30:38.006603 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice/crio-f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice\": RecentStats: unable to find data in memory cache]" Dec 03 17:30:46 crc kubenswrapper[4787]: I1203 17:30:46.378671 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-86b8bcb77d-fd8v5" Dec 03 17:30:48 crc kubenswrapper[4787]: E1203 17:30:48.163701 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice/crio-f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice\": RecentStats: unable to find data in memory cache]" Dec 03 17:30:48 crc kubenswrapper[4787]: I1203 17:30:48.990350 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:30:48 crc kubenswrapper[4787]: I1203 17:30:48.990446 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:30:48 crc kubenswrapper[4787]: I1203 17:30:48.990551 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:30:48 crc kubenswrapper[4787]: I1203 17:30:48.991615 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"864b789057addccde661186743b5963a0f8e318c2a7a639d2b92c72f83cc12f6"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:30:48 crc kubenswrapper[4787]: I1203 17:30:48.991752 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://864b789057addccde661186743b5963a0f8e318c2a7a639d2b92c72f83cc12f6" gracePeriod=600 Dec 03 17:30:49 crc kubenswrapper[4787]: I1203 17:30:49.349594 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="864b789057addccde661186743b5963a0f8e318c2a7a639d2b92c72f83cc12f6" exitCode=0 Dec 03 17:30:49 crc kubenswrapper[4787]: I1203 17:30:49.349694 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"864b789057addccde661186743b5963a0f8e318c2a7a639d2b92c72f83cc12f6"} Dec 03 17:30:49 crc kubenswrapper[4787]: I1203 17:30:49.349896 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"93e6b55e4dd0c75878fc7a5cbe5a33dad37bb97aa5cbfc657a6c1fd769073445"} Dec 03 17:30:49 crc kubenswrapper[4787]: I1203 17:30:49.349926 4787 scope.go:117] "RemoveContainer" containerID="49596b58389f8b6bbf73d259649cd03b2e93ae7e21af7479d7bc4161137e279b" Dec 03 17:30:58 crc kubenswrapper[4787]: E1203 17:30:58.315359 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda5b1845b_dcdb_4fb2_983a_4018aa964656.slice/crio-f39782a5b2b325346db7f4899dfbc9283456329ca3535b44a66f2d272d6337a7\": RecentStats: unable to find data in memory cache]" Dec 03 17:31:05 crc kubenswrapper[4787]: I1203 17:31:05.939888 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-9b64dcbdc-qzmjb" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.782452 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-pgt2s"] Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.785776 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.810989 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-6kplw" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.828661 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2"] Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.834617 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.834868 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.834867 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.843958 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.847463 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2"] Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.865783 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-frr-sockets\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.865866 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-reloader\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.865886 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-frr-conf\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.865919 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlp8q\" (UniqueName: \"kubernetes.io/projected/37601185-c3bd-4614-b119-05f1b07f2875-kube-api-access-xlp8q\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.865942 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-metrics\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.866079 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/37601185-c3bd-4614-b119-05f1b07f2875-frr-startup\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.866096 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37601185-c3bd-4614-b119-05f1b07f2875-metrics-certs\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.887192 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-sfcqc"] Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.888403 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-sfcqc" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.891691 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.891873 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-jz2nv" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.891945 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.891984 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.911740 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-6rp2q"] Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.913145 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.915877 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.926856 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-6rp2q"] Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968062 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w99f\" (UniqueName: \"kubernetes.io/projected/d571a3dc-bd38-4881-857f-e7986b1d90af-kube-api-access-9w99f\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968129 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nq4fg\" (UniqueName: \"kubernetes.io/projected/e1c08eda-0859-42f8-9ca2-f1c4df77038c-kube-api-access-nq4fg\") pod \"controller-f8648f98b-6rp2q\" (UID: \"e1c08eda-0859-42f8-9ca2-f1c4df77038c\") " pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968177 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/37601185-c3bd-4614-b119-05f1b07f2875-frr-startup\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968203 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37601185-c3bd-4614-b119-05f1b07f2875-metrics-certs\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968233 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-frr-sockets\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968262 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwcq7\" (UniqueName: \"kubernetes.io/projected/55b2904e-d04c-43ef-b054-0ef8636a2316-kube-api-access-pwcq7\") pod \"frr-k8s-webhook-server-7fcb986d4-kqsb2\" (UID: \"55b2904e-d04c-43ef-b054-0ef8636a2316\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968287 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-metrics-certs\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968323 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d571a3dc-bd38-4881-857f-e7986b1d90af-metallb-excludel2\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968367 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-reloader\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968391 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-frr-conf\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968429 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55b2904e-d04c-43ef-b054-0ef8636a2316-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-kqsb2\" (UID: \"55b2904e-d04c-43ef-b054-0ef8636a2316\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968467 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlp8q\" (UniqueName: \"kubernetes.io/projected/37601185-c3bd-4614-b119-05f1b07f2875-kube-api-access-xlp8q\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968499 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-metrics\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968541 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1c08eda-0859-42f8-9ca2-f1c4df77038c-metrics-certs\") pod \"controller-f8648f98b-6rp2q\" (UID: \"e1c08eda-0859-42f8-9ca2-f1c4df77038c\") " pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968568 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-memberlist\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.968595 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e1c08eda-0859-42f8-9ca2-f1c4df77038c-cert\") pod \"controller-f8648f98b-6rp2q\" (UID: \"e1c08eda-0859-42f8-9ca2-f1c4df77038c\") " pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.969721 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/37601185-c3bd-4614-b119-05f1b07f2875-frr-startup\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: E1203 17:31:06.969831 4787 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 03 17:31:06 crc kubenswrapper[4787]: E1203 17:31:06.969915 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/37601185-c3bd-4614-b119-05f1b07f2875-metrics-certs podName:37601185-c3bd-4614-b119-05f1b07f2875 nodeName:}" failed. No retries permitted until 2025-12-03 17:31:07.469873977 +0000 UTC m=+1084.287344936 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/37601185-c3bd-4614-b119-05f1b07f2875-metrics-certs") pod "frr-k8s-pgt2s" (UID: "37601185-c3bd-4614-b119-05f1b07f2875") : secret "frr-k8s-certs-secret" not found Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.970518 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-frr-sockets\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.970788 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-reloader\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.971296 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-metrics\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.971550 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/37601185-c3bd-4614-b119-05f1b07f2875-frr-conf\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:06 crc kubenswrapper[4787]: I1203 17:31:06.990237 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlp8q\" (UniqueName: \"kubernetes.io/projected/37601185-c3bd-4614-b119-05f1b07f2875-kube-api-access-xlp8q\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.069864 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1c08eda-0859-42f8-9ca2-f1c4df77038c-metrics-certs\") pod \"controller-f8648f98b-6rp2q\" (UID: \"e1c08eda-0859-42f8-9ca2-f1c4df77038c\") " pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.069920 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-memberlist\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.069943 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e1c08eda-0859-42f8-9ca2-f1c4df77038c-cert\") pod \"controller-f8648f98b-6rp2q\" (UID: \"e1c08eda-0859-42f8-9ca2-f1c4df77038c\") " pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.069985 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w99f\" (UniqueName: \"kubernetes.io/projected/d571a3dc-bd38-4881-857f-e7986b1d90af-kube-api-access-9w99f\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.070006 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nq4fg\" (UniqueName: \"kubernetes.io/projected/e1c08eda-0859-42f8-9ca2-f1c4df77038c-kube-api-access-nq4fg\") pod \"controller-f8648f98b-6rp2q\" (UID: \"e1c08eda-0859-42f8-9ca2-f1c4df77038c\") " pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.070056 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwcq7\" (UniqueName: \"kubernetes.io/projected/55b2904e-d04c-43ef-b054-0ef8636a2316-kube-api-access-pwcq7\") pod \"frr-k8s-webhook-server-7fcb986d4-kqsb2\" (UID: \"55b2904e-d04c-43ef-b054-0ef8636a2316\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.070075 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-metrics-certs\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.070097 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d571a3dc-bd38-4881-857f-e7986b1d90af-metallb-excludel2\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.070125 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55b2904e-d04c-43ef-b054-0ef8636a2316-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-kqsb2\" (UID: \"55b2904e-d04c-43ef-b054-0ef8636a2316\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" Dec 03 17:31:07 crc kubenswrapper[4787]: E1203 17:31:07.070209 4787 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 03 17:31:07 crc kubenswrapper[4787]: E1203 17:31:07.070256 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/55b2904e-d04c-43ef-b054-0ef8636a2316-cert podName:55b2904e-d04c-43ef-b054-0ef8636a2316 nodeName:}" failed. No retries permitted until 2025-12-03 17:31:07.570241939 +0000 UTC m=+1084.387712898 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/55b2904e-d04c-43ef-b054-0ef8636a2316-cert") pod "frr-k8s-webhook-server-7fcb986d4-kqsb2" (UID: "55b2904e-d04c-43ef-b054-0ef8636a2316") : secret "frr-k8s-webhook-server-cert" not found Dec 03 17:31:07 crc kubenswrapper[4787]: E1203 17:31:07.070451 4787 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 17:31:07 crc kubenswrapper[4787]: E1203 17:31:07.070474 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-memberlist podName:d571a3dc-bd38-4881-857f-e7986b1d90af nodeName:}" failed. No retries permitted until 2025-12-03 17:31:07.570466495 +0000 UTC m=+1084.387937454 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-memberlist") pod "speaker-sfcqc" (UID: "d571a3dc-bd38-4881-857f-e7986b1d90af") : secret "metallb-memberlist" not found Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.071202 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d571a3dc-bd38-4881-857f-e7986b1d90af-metallb-excludel2\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.073976 4787 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.078494 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e1c08eda-0859-42f8-9ca2-f1c4df77038c-metrics-certs\") pod \"controller-f8648f98b-6rp2q\" (UID: \"e1c08eda-0859-42f8-9ca2-f1c4df77038c\") " pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.090798 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-metrics-certs\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.092560 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w99f\" (UniqueName: \"kubernetes.io/projected/d571a3dc-bd38-4881-857f-e7986b1d90af-kube-api-access-9w99f\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.092844 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/e1c08eda-0859-42f8-9ca2-f1c4df77038c-cert\") pod \"controller-f8648f98b-6rp2q\" (UID: \"e1c08eda-0859-42f8-9ca2-f1c4df77038c\") " pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.095368 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwcq7\" (UniqueName: \"kubernetes.io/projected/55b2904e-d04c-43ef-b054-0ef8636a2316-kube-api-access-pwcq7\") pod \"frr-k8s-webhook-server-7fcb986d4-kqsb2\" (UID: \"55b2904e-d04c-43ef-b054-0ef8636a2316\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.095993 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nq4fg\" (UniqueName: \"kubernetes.io/projected/e1c08eda-0859-42f8-9ca2-f1c4df77038c-kube-api-access-nq4fg\") pod \"controller-f8648f98b-6rp2q\" (UID: \"e1c08eda-0859-42f8-9ca2-f1c4df77038c\") " pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.231844 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.478029 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37601185-c3bd-4614-b119-05f1b07f2875-metrics-certs\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.483595 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/37601185-c3bd-4614-b119-05f1b07f2875-metrics-certs\") pod \"frr-k8s-pgt2s\" (UID: \"37601185-c3bd-4614-b119-05f1b07f2875\") " pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.579942 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55b2904e-d04c-43ef-b054-0ef8636a2316-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-kqsb2\" (UID: \"55b2904e-d04c-43ef-b054-0ef8636a2316\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.580045 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-memberlist\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:07 crc kubenswrapper[4787]: E1203 17:31:07.580264 4787 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 03 17:31:07 crc kubenswrapper[4787]: E1203 17:31:07.580328 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-memberlist podName:d571a3dc-bd38-4881-857f-e7986b1d90af nodeName:}" failed. No retries permitted until 2025-12-03 17:31:08.580308175 +0000 UTC m=+1085.397779134 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-memberlist") pod "speaker-sfcqc" (UID: "d571a3dc-bd38-4881-857f-e7986b1d90af") : secret "metallb-memberlist" not found Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.583361 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/55b2904e-d04c-43ef-b054-0ef8636a2316-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-kqsb2\" (UID: \"55b2904e-d04c-43ef-b054-0ef8636a2316\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.647828 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-6rp2q"] Dec 03 17:31:07 crc kubenswrapper[4787]: W1203 17:31:07.654190 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1c08eda_0859_42f8_9ca2_f1c4df77038c.slice/crio-a47ebf1e0251e7eb0e927d12bc9ddfd1c0732fd898ccee14642620a399d2b45a WatchSource:0}: Error finding container a47ebf1e0251e7eb0e927d12bc9ddfd1c0732fd898ccee14642620a399d2b45a: Status 404 returned error can't find the container with id a47ebf1e0251e7eb0e927d12bc9ddfd1c0732fd898ccee14642620a399d2b45a Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.736615 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:07 crc kubenswrapper[4787]: I1203 17:31:07.765563 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" Dec 03 17:31:08 crc kubenswrapper[4787]: I1203 17:31:08.211515 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2"] Dec 03 17:31:08 crc kubenswrapper[4787]: I1203 17:31:08.488320 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" event={"ID":"55b2904e-d04c-43ef-b054-0ef8636a2316","Type":"ContainerStarted","Data":"fd01b91448b888f1ae0fea18bfaa5715f13885b9b46c559545ffe681cc0ea1cb"} Dec 03 17:31:08 crc kubenswrapper[4787]: I1203 17:31:08.489402 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgt2s" event={"ID":"37601185-c3bd-4614-b119-05f1b07f2875","Type":"ContainerStarted","Data":"61822ce12bbc3aa7eac96b13eb1eb25b0f6bc55cae6ddbc93f8662999019a904"} Dec 03 17:31:08 crc kubenswrapper[4787]: I1203 17:31:08.491241 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6rp2q" event={"ID":"e1c08eda-0859-42f8-9ca2-f1c4df77038c","Type":"ContainerStarted","Data":"885c6b35306aab53be5538ef4d77351be02f1656ed4db2f63eb03ceea15438c1"} Dec 03 17:31:08 crc kubenswrapper[4787]: I1203 17:31:08.491284 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6rp2q" event={"ID":"e1c08eda-0859-42f8-9ca2-f1c4df77038c","Type":"ContainerStarted","Data":"dcf84c4d631831e3f444fcf655be940b482004810d0e316b294b46f3092424a7"} Dec 03 17:31:08 crc kubenswrapper[4787]: I1203 17:31:08.491301 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-6rp2q" event={"ID":"e1c08eda-0859-42f8-9ca2-f1c4df77038c","Type":"ContainerStarted","Data":"a47ebf1e0251e7eb0e927d12bc9ddfd1c0732fd898ccee14642620a399d2b45a"} Dec 03 17:31:08 crc kubenswrapper[4787]: I1203 17:31:08.491419 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:08 crc kubenswrapper[4787]: I1203 17:31:08.510217 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-6rp2q" podStartSLOduration=2.510176115 podStartE2EDuration="2.510176115s" podCreationTimestamp="2025-12-03 17:31:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:31:08.507588846 +0000 UTC m=+1085.325059825" watchObservedRunningTime="2025-12-03 17:31:08.510176115 +0000 UTC m=+1085.327647074" Dec 03 17:31:08 crc kubenswrapper[4787]: I1203 17:31:08.596773 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-memberlist\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:08 crc kubenswrapper[4787]: I1203 17:31:08.602619 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d571a3dc-bd38-4881-857f-e7986b1d90af-memberlist\") pod \"speaker-sfcqc\" (UID: \"d571a3dc-bd38-4881-857f-e7986b1d90af\") " pod="metallb-system/speaker-sfcqc" Dec 03 17:31:08 crc kubenswrapper[4787]: I1203 17:31:08.706205 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-sfcqc" Dec 03 17:31:09 crc kubenswrapper[4787]: I1203 17:31:09.520771 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sfcqc" event={"ID":"d571a3dc-bd38-4881-857f-e7986b1d90af","Type":"ContainerStarted","Data":"8baffca87074761313ea2d6691ee477a7b87524c384273960b87266f955cc861"} Dec 03 17:31:09 crc kubenswrapper[4787]: I1203 17:31:09.521102 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sfcqc" event={"ID":"d571a3dc-bd38-4881-857f-e7986b1d90af","Type":"ContainerStarted","Data":"6be51dfafcde4999835bdc327a9fe9c29ec57d923f1f9994312116fafa0ba7f6"} Dec 03 17:31:10 crc kubenswrapper[4787]: I1203 17:31:10.547970 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-sfcqc" event={"ID":"d571a3dc-bd38-4881-857f-e7986b1d90af","Type":"ContainerStarted","Data":"b95cdc3c6387c003e3d00df24f5be02e8ac543ec89730ee01b32b3c5e7a0cd7e"} Dec 03 17:31:10 crc kubenswrapper[4787]: I1203 17:31:10.548348 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-sfcqc" Dec 03 17:31:10 crc kubenswrapper[4787]: I1203 17:31:10.586062 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-sfcqc" podStartSLOduration=4.586000623 podStartE2EDuration="4.586000623s" podCreationTimestamp="2025-12-03 17:31:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:31:10.57797627 +0000 UTC m=+1087.395447249" watchObservedRunningTime="2025-12-03 17:31:10.586000623 +0000 UTC m=+1087.403471582" Dec 03 17:31:17 crc kubenswrapper[4787]: I1203 17:31:17.237287 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-6rp2q" Dec 03 17:31:18 crc kubenswrapper[4787]: I1203 17:31:18.608395 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" event={"ID":"55b2904e-d04c-43ef-b054-0ef8636a2316","Type":"ContainerStarted","Data":"c6f36fc82c5e13c8bf9d5b50f1b674bfcaf190cbabbfb69397e5908b814bb77e"} Dec 03 17:31:18 crc kubenswrapper[4787]: I1203 17:31:18.610566 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" Dec 03 17:31:18 crc kubenswrapper[4787]: I1203 17:31:18.623549 4787 generic.go:334] "Generic (PLEG): container finished" podID="37601185-c3bd-4614-b119-05f1b07f2875" containerID="7f5a577b90d94a4c24099fd89ddcba0589f5736eb6a365427afadf7e3a0243cb" exitCode=0 Dec 03 17:31:18 crc kubenswrapper[4787]: I1203 17:31:18.623613 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgt2s" event={"ID":"37601185-c3bd-4614-b119-05f1b07f2875","Type":"ContainerDied","Data":"7f5a577b90d94a4c24099fd89ddcba0589f5736eb6a365427afadf7e3a0243cb"} Dec 03 17:31:18 crc kubenswrapper[4787]: I1203 17:31:18.675250 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" podStartSLOduration=3.700214944 podStartE2EDuration="12.675228353s" podCreationTimestamp="2025-12-03 17:31:06 +0000 UTC" firstStartedPulling="2025-12-03 17:31:08.213156538 +0000 UTC m=+1085.030627497" lastFinishedPulling="2025-12-03 17:31:17.188169947 +0000 UTC m=+1094.005640906" observedRunningTime="2025-12-03 17:31:18.644380865 +0000 UTC m=+1095.461851844" watchObservedRunningTime="2025-12-03 17:31:18.675228353 +0000 UTC m=+1095.492699312" Dec 03 17:31:19 crc kubenswrapper[4787]: I1203 17:31:19.632212 4787 generic.go:334] "Generic (PLEG): container finished" podID="37601185-c3bd-4614-b119-05f1b07f2875" containerID="977433f24782c98b2d2942339fad2e6ea2f6125d98d0c8bb09a7aaa716bfa7fc" exitCode=0 Dec 03 17:31:19 crc kubenswrapper[4787]: I1203 17:31:19.632336 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgt2s" event={"ID":"37601185-c3bd-4614-b119-05f1b07f2875","Type":"ContainerDied","Data":"977433f24782c98b2d2942339fad2e6ea2f6125d98d0c8bb09a7aaa716bfa7fc"} Dec 03 17:31:20 crc kubenswrapper[4787]: I1203 17:31:20.641260 4787 generic.go:334] "Generic (PLEG): container finished" podID="37601185-c3bd-4614-b119-05f1b07f2875" containerID="495ca12e3c0652149e224911c923e7a74dd622a8c110a2aae9746c7988ab07a6" exitCode=0 Dec 03 17:31:20 crc kubenswrapper[4787]: I1203 17:31:20.641290 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgt2s" event={"ID":"37601185-c3bd-4614-b119-05f1b07f2875","Type":"ContainerDied","Data":"495ca12e3c0652149e224911c923e7a74dd622a8c110a2aae9746c7988ab07a6"} Dec 03 17:31:21 crc kubenswrapper[4787]: I1203 17:31:21.650511 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgt2s" event={"ID":"37601185-c3bd-4614-b119-05f1b07f2875","Type":"ContainerStarted","Data":"d9a561dca24c1e0323e755ea2593a5c1b9a537fcec8026eeafdfe35352e161a5"} Dec 03 17:31:21 crc kubenswrapper[4787]: I1203 17:31:21.651667 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgt2s" event={"ID":"37601185-c3bd-4614-b119-05f1b07f2875","Type":"ContainerStarted","Data":"025a2bf95351e29dd42b944f04a6c0274ecc88a1f9679d556ad83f9343d6115a"} Dec 03 17:31:21 crc kubenswrapper[4787]: I1203 17:31:21.651782 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgt2s" event={"ID":"37601185-c3bd-4614-b119-05f1b07f2875","Type":"ContainerStarted","Data":"1a967a295ce4d37ef0792132db57e795498fc82abd83b8726e3a3c898b9324e7"} Dec 03 17:31:21 crc kubenswrapper[4787]: I1203 17:31:21.651862 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgt2s" event={"ID":"37601185-c3bd-4614-b119-05f1b07f2875","Type":"ContainerStarted","Data":"65d7f6d4375f947887a1f50e8e7eccc89ba5868274096cf4e93cd4e66472b233"} Dec 03 17:31:21 crc kubenswrapper[4787]: I1203 17:31:21.651947 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgt2s" event={"ID":"37601185-c3bd-4614-b119-05f1b07f2875","Type":"ContainerStarted","Data":"2bcaba19ae4bab6b624e49a322178d3a02e217d2440bb01b9a105313a3dcbb03"} Dec 03 17:31:22 crc kubenswrapper[4787]: I1203 17:31:22.662725 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-pgt2s" event={"ID":"37601185-c3bd-4614-b119-05f1b07f2875","Type":"ContainerStarted","Data":"7309c37e9ea8ea79edaf1afde7f7968ee57088db02036a70418c15efcee51bfc"} Dec 03 17:31:22 crc kubenswrapper[4787]: I1203 17:31:22.662993 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:22 crc kubenswrapper[4787]: I1203 17:31:22.693066 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-pgt2s" podStartSLOduration=7.348908314 podStartE2EDuration="16.693046262s" podCreationTimestamp="2025-12-03 17:31:06 +0000 UTC" firstStartedPulling="2025-12-03 17:31:07.867668996 +0000 UTC m=+1084.685139955" lastFinishedPulling="2025-12-03 17:31:17.211806944 +0000 UTC m=+1094.029277903" observedRunningTime="2025-12-03 17:31:22.687837694 +0000 UTC m=+1099.505308663" watchObservedRunningTime="2025-12-03 17:31:22.693046262 +0000 UTC m=+1099.510517221" Dec 03 17:31:22 crc kubenswrapper[4787]: I1203 17:31:22.736877 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:22 crc kubenswrapper[4787]: I1203 17:31:22.789276 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:27 crc kubenswrapper[4787]: I1203 17:31:27.774841 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-kqsb2" Dec 03 17:31:28 crc kubenswrapper[4787]: I1203 17:31:28.709769 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-sfcqc" Dec 03 17:31:31 crc kubenswrapper[4787]: I1203 17:31:31.808844 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-p94x6"] Dec 03 17:31:31 crc kubenswrapper[4787]: I1203 17:31:31.811096 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-p94x6" Dec 03 17:31:31 crc kubenswrapper[4787]: I1203 17:31:31.825435 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 03 17:31:31 crc kubenswrapper[4787]: I1203 17:31:31.826631 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 03 17:31:31 crc kubenswrapper[4787]: I1203 17:31:31.826837 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-9xsp6" Dec 03 17:31:31 crc kubenswrapper[4787]: I1203 17:31:31.844536 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-p94x6"] Dec 03 17:31:31 crc kubenswrapper[4787]: I1203 17:31:31.896085 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ddvqm\" (UniqueName: \"kubernetes.io/projected/20a259a7-6577-484d-91a8-7c47546eff90-kube-api-access-ddvqm\") pod \"openstack-operator-index-p94x6\" (UID: \"20a259a7-6577-484d-91a8-7c47546eff90\") " pod="openstack-operators/openstack-operator-index-p94x6" Dec 03 17:31:31 crc kubenswrapper[4787]: I1203 17:31:31.997725 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ddvqm\" (UniqueName: \"kubernetes.io/projected/20a259a7-6577-484d-91a8-7c47546eff90-kube-api-access-ddvqm\") pod \"openstack-operator-index-p94x6\" (UID: \"20a259a7-6577-484d-91a8-7c47546eff90\") " pod="openstack-operators/openstack-operator-index-p94x6" Dec 03 17:31:32 crc kubenswrapper[4787]: I1203 17:31:32.019850 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ddvqm\" (UniqueName: \"kubernetes.io/projected/20a259a7-6577-484d-91a8-7c47546eff90-kube-api-access-ddvqm\") pod \"openstack-operator-index-p94x6\" (UID: \"20a259a7-6577-484d-91a8-7c47546eff90\") " pod="openstack-operators/openstack-operator-index-p94x6" Dec 03 17:31:32 crc kubenswrapper[4787]: I1203 17:31:32.147283 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-p94x6" Dec 03 17:31:32 crc kubenswrapper[4787]: I1203 17:31:32.591893 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-p94x6"] Dec 03 17:31:32 crc kubenswrapper[4787]: W1203 17:31:32.602601 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod20a259a7_6577_484d_91a8_7c47546eff90.slice/crio-08b115aae468d5f2ccc6ba619ca528644102508103525f0c25980c7ab78e05f8 WatchSource:0}: Error finding container 08b115aae468d5f2ccc6ba619ca528644102508103525f0c25980c7ab78e05f8: Status 404 returned error can't find the container with id 08b115aae468d5f2ccc6ba619ca528644102508103525f0c25980c7ab78e05f8 Dec 03 17:31:32 crc kubenswrapper[4787]: I1203 17:31:32.742547 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-p94x6" event={"ID":"20a259a7-6577-484d-91a8-7c47546eff90","Type":"ContainerStarted","Data":"08b115aae468d5f2ccc6ba619ca528644102508103525f0c25980c7ab78e05f8"} Dec 03 17:31:34 crc kubenswrapper[4787]: I1203 17:31:34.176096 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-p94x6"] Dec 03 17:31:34 crc kubenswrapper[4787]: I1203 17:31:34.579777 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-t5bxl"] Dec 03 17:31:34 crc kubenswrapper[4787]: I1203 17:31:34.583047 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t5bxl" Dec 03 17:31:34 crc kubenswrapper[4787]: I1203 17:31:34.602968 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t5bxl"] Dec 03 17:31:34 crc kubenswrapper[4787]: I1203 17:31:34.641776 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thb5q\" (UniqueName: \"kubernetes.io/projected/e5effab2-18ac-4ad1-8d08-618f20509a85-kube-api-access-thb5q\") pod \"openstack-operator-index-t5bxl\" (UID: \"e5effab2-18ac-4ad1-8d08-618f20509a85\") " pod="openstack-operators/openstack-operator-index-t5bxl" Dec 03 17:31:34 crc kubenswrapper[4787]: I1203 17:31:34.743638 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thb5q\" (UniqueName: \"kubernetes.io/projected/e5effab2-18ac-4ad1-8d08-618f20509a85-kube-api-access-thb5q\") pod \"openstack-operator-index-t5bxl\" (UID: \"e5effab2-18ac-4ad1-8d08-618f20509a85\") " pod="openstack-operators/openstack-operator-index-t5bxl" Dec 03 17:31:34 crc kubenswrapper[4787]: I1203 17:31:34.767108 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thb5q\" (UniqueName: \"kubernetes.io/projected/e5effab2-18ac-4ad1-8d08-618f20509a85-kube-api-access-thb5q\") pod \"openstack-operator-index-t5bxl\" (UID: \"e5effab2-18ac-4ad1-8d08-618f20509a85\") " pod="openstack-operators/openstack-operator-index-t5bxl" Dec 03 17:31:34 crc kubenswrapper[4787]: I1203 17:31:34.914266 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-t5bxl" Dec 03 17:31:37 crc kubenswrapper[4787]: I1203 17:31:37.234548 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-t5bxl"] Dec 03 17:31:37 crc kubenswrapper[4787]: I1203 17:31:37.742691 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-pgt2s" Dec 03 17:31:37 crc kubenswrapper[4787]: I1203 17:31:37.793175 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t5bxl" event={"ID":"e5effab2-18ac-4ad1-8d08-618f20509a85","Type":"ContainerStarted","Data":"cc52b4b57cfcdb00efa153371181c3bcefbeeba74c12bd673a6ad78d03d3a72c"} Dec 03 17:31:37 crc kubenswrapper[4787]: I1203 17:31:37.793525 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-t5bxl" event={"ID":"e5effab2-18ac-4ad1-8d08-618f20509a85","Type":"ContainerStarted","Data":"1528736e234e77eae116b6007be128f8ed125670748eabc13c733cc49b3d977c"} Dec 03 17:31:37 crc kubenswrapper[4787]: I1203 17:31:37.794847 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-p94x6" event={"ID":"20a259a7-6577-484d-91a8-7c47546eff90","Type":"ContainerStarted","Data":"e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2"} Dec 03 17:31:37 crc kubenswrapper[4787]: I1203 17:31:37.795007 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-p94x6" podUID="20a259a7-6577-484d-91a8-7c47546eff90" containerName="registry-server" containerID="cri-o://e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2" gracePeriod=2 Dec 03 17:31:37 crc kubenswrapper[4787]: I1203 17:31:37.811440 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-t5bxl" podStartSLOduration=3.755203377 podStartE2EDuration="3.811416318s" podCreationTimestamp="2025-12-03 17:31:34 +0000 UTC" firstStartedPulling="2025-12-03 17:31:37.244671368 +0000 UTC m=+1114.062142337" lastFinishedPulling="2025-12-03 17:31:37.300884319 +0000 UTC m=+1114.118355278" observedRunningTime="2025-12-03 17:31:37.810822602 +0000 UTC m=+1114.628293571" watchObservedRunningTime="2025-12-03 17:31:37.811416318 +0000 UTC m=+1114.628887277" Dec 03 17:31:37 crc kubenswrapper[4787]: I1203 17:31:37.838662 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-p94x6" podStartSLOduration=2.654699366 podStartE2EDuration="6.83863933s" podCreationTimestamp="2025-12-03 17:31:31 +0000 UTC" firstStartedPulling="2025-12-03 17:31:32.605214414 +0000 UTC m=+1109.422685373" lastFinishedPulling="2025-12-03 17:31:36.789154378 +0000 UTC m=+1113.606625337" observedRunningTime="2025-12-03 17:31:37.83338332 +0000 UTC m=+1114.650854289" watchObservedRunningTime="2025-12-03 17:31:37.83863933 +0000 UTC m=+1114.656110289" Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.219638 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-p94x6" Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.304793 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ddvqm\" (UniqueName: \"kubernetes.io/projected/20a259a7-6577-484d-91a8-7c47546eff90-kube-api-access-ddvqm\") pod \"20a259a7-6577-484d-91a8-7c47546eff90\" (UID: \"20a259a7-6577-484d-91a8-7c47546eff90\") " Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.315256 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20a259a7-6577-484d-91a8-7c47546eff90-kube-api-access-ddvqm" (OuterVolumeSpecName: "kube-api-access-ddvqm") pod "20a259a7-6577-484d-91a8-7c47546eff90" (UID: "20a259a7-6577-484d-91a8-7c47546eff90"). InnerVolumeSpecName "kube-api-access-ddvqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.407422 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ddvqm\" (UniqueName: \"kubernetes.io/projected/20a259a7-6577-484d-91a8-7c47546eff90-kube-api-access-ddvqm\") on node \"crc\" DevicePath \"\"" Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.822065 4787 generic.go:334] "Generic (PLEG): container finished" podID="20a259a7-6577-484d-91a8-7c47546eff90" containerID="e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2" exitCode=0 Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.822138 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-p94x6" event={"ID":"20a259a7-6577-484d-91a8-7c47546eff90","Type":"ContainerDied","Data":"e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2"} Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.822505 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-p94x6" event={"ID":"20a259a7-6577-484d-91a8-7c47546eff90","Type":"ContainerDied","Data":"08b115aae468d5f2ccc6ba619ca528644102508103525f0c25980c7ab78e05f8"} Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.822162 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-p94x6" Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.822569 4787 scope.go:117] "RemoveContainer" containerID="e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2" Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.858283 4787 scope.go:117] "RemoveContainer" containerID="e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2" Dec 03 17:31:38 crc kubenswrapper[4787]: E1203 17:31:38.865510 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2\": container with ID starting with e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2 not found: ID does not exist" containerID="e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2" Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.865693 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2"} err="failed to get container status \"e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2\": rpc error: code = NotFound desc = could not find container \"e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2\": container with ID starting with e8a6ea64d9e732a1aa0384630208b40aa2c865c8d454df6830bfdef7992789c2 not found: ID does not exist" Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.865806 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-p94x6"] Dec 03 17:31:38 crc kubenswrapper[4787]: I1203 17:31:38.873462 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-p94x6"] Dec 03 17:31:39 crc kubenswrapper[4787]: I1203 17:31:39.774990 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20a259a7-6577-484d-91a8-7c47546eff90" path="/var/lib/kubelet/pods/20a259a7-6577-484d-91a8-7c47546eff90/volumes" Dec 03 17:31:44 crc kubenswrapper[4787]: I1203 17:31:44.914627 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-t5bxl" Dec 03 17:31:44 crc kubenswrapper[4787]: I1203 17:31:44.915080 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-t5bxl" Dec 03 17:31:44 crc kubenswrapper[4787]: I1203 17:31:44.945235 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-t5bxl" Dec 03 17:31:45 crc kubenswrapper[4787]: I1203 17:31:45.909611 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-t5bxl" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.628690 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz"] Dec 03 17:31:48 crc kubenswrapper[4787]: E1203 17:31:48.629293 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="20a259a7-6577-484d-91a8-7c47546eff90" containerName="registry-server" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.629306 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="20a259a7-6577-484d-91a8-7c47546eff90" containerName="registry-server" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.629457 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="20a259a7-6577-484d-91a8-7c47546eff90" containerName="registry-server" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.630617 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.634615 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-58rbm" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.646482 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz"] Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.785357 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-bundle\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.785512 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-util\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.785561 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlqr9\" (UniqueName: \"kubernetes.io/projected/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-kube-api-access-qlqr9\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.886808 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlqr9\" (UniqueName: \"kubernetes.io/projected/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-kube-api-access-qlqr9\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.886879 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-bundle\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.886940 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-util\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.887391 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-util\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.887484 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-bundle\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.930648 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlqr9\" (UniqueName: \"kubernetes.io/projected/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-kube-api-access-qlqr9\") pod \"f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:48 crc kubenswrapper[4787]: I1203 17:31:48.957532 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:49 crc kubenswrapper[4787]: I1203 17:31:49.397779 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz"] Dec 03 17:31:49 crc kubenswrapper[4787]: I1203 17:31:49.900529 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" event={"ID":"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be","Type":"ContainerStarted","Data":"9b5e86f10ea0e9056a6e7a0cbcd7d4cadf21cf49f1e3b713f8ca773a5541160d"} Dec 03 17:31:49 crc kubenswrapper[4787]: I1203 17:31:49.900875 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" event={"ID":"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be","Type":"ContainerStarted","Data":"2d4152011f6fc74769721f3282e954ae2f36101b20bb1a2cf2cb6e9c574c85de"} Dec 03 17:31:50 crc kubenswrapper[4787]: I1203 17:31:50.909840 4787 generic.go:334] "Generic (PLEG): container finished" podID="76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" containerID="9b5e86f10ea0e9056a6e7a0cbcd7d4cadf21cf49f1e3b713f8ca773a5541160d" exitCode=0 Dec 03 17:31:50 crc kubenswrapper[4787]: I1203 17:31:50.909910 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" event={"ID":"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be","Type":"ContainerDied","Data":"9b5e86f10ea0e9056a6e7a0cbcd7d4cadf21cf49f1e3b713f8ca773a5541160d"} Dec 03 17:31:52 crc kubenswrapper[4787]: I1203 17:31:52.924094 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" event={"ID":"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be","Type":"ContainerStarted","Data":"153ee7fdcc3be16b991c89c5b562c78ae26c94e4a47add77ac1118e63ffe1c23"} Dec 03 17:31:53 crc kubenswrapper[4787]: I1203 17:31:53.934193 4787 generic.go:334] "Generic (PLEG): container finished" podID="76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" containerID="153ee7fdcc3be16b991c89c5b562c78ae26c94e4a47add77ac1118e63ffe1c23" exitCode=0 Dec 03 17:31:53 crc kubenswrapper[4787]: I1203 17:31:53.934264 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" event={"ID":"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be","Type":"ContainerDied","Data":"153ee7fdcc3be16b991c89c5b562c78ae26c94e4a47add77ac1118e63ffe1c23"} Dec 03 17:31:54 crc kubenswrapper[4787]: I1203 17:31:54.946683 4787 generic.go:334] "Generic (PLEG): container finished" podID="76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" containerID="1b493858cb40e8c7fd68d65bb35beee3b46e389fedeb2c16ce3d930d2ecdf8a2" exitCode=0 Dec 03 17:31:54 crc kubenswrapper[4787]: I1203 17:31:54.946771 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" event={"ID":"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be","Type":"ContainerDied","Data":"1b493858cb40e8c7fd68d65bb35beee3b46e389fedeb2c16ce3d930d2ecdf8a2"} Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.325287 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.511855 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-bundle\") pod \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.512000 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-util\") pod \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.512175 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qlqr9\" (UniqueName: \"kubernetes.io/projected/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-kube-api-access-qlqr9\") pod \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\" (UID: \"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be\") " Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.513658 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-bundle" (OuterVolumeSpecName: "bundle") pod "76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" (UID: "76ec0f29-6ea6-42a5-a4cc-43d5a8b928be"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.519369 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-kube-api-access-qlqr9" (OuterVolumeSpecName: "kube-api-access-qlqr9") pod "76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" (UID: "76ec0f29-6ea6-42a5-a4cc-43d5a8b928be"). InnerVolumeSpecName "kube-api-access-qlqr9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.526808 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-util" (OuterVolumeSpecName: "util") pod "76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" (UID: "76ec0f29-6ea6-42a5-a4cc-43d5a8b928be"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.614077 4787 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.614109 4787 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-util\") on node \"crc\" DevicePath \"\"" Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.614119 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qlqr9\" (UniqueName: \"kubernetes.io/projected/76ec0f29-6ea6-42a5-a4cc-43d5a8b928be-kube-api-access-qlqr9\") on node \"crc\" DevicePath \"\"" Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.963201 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" event={"ID":"76ec0f29-6ea6-42a5-a4cc-43d5a8b928be","Type":"ContainerDied","Data":"2d4152011f6fc74769721f3282e954ae2f36101b20bb1a2cf2cb6e9c574c85de"} Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.963253 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2d4152011f6fc74769721f3282e954ae2f36101b20bb1a2cf2cb6e9c574c85de" Dec 03 17:31:56 crc kubenswrapper[4787]: I1203 17:31:56.963274 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz" Dec 03 17:32:01 crc kubenswrapper[4787]: I1203 17:32:01.937917 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758"] Dec 03 17:32:01 crc kubenswrapper[4787]: E1203 17:32:01.939455 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" containerName="util" Dec 03 17:32:01 crc kubenswrapper[4787]: I1203 17:32:01.939483 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" containerName="util" Dec 03 17:32:01 crc kubenswrapper[4787]: E1203 17:32:01.939535 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" containerName="extract" Dec 03 17:32:01 crc kubenswrapper[4787]: I1203 17:32:01.939575 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" containerName="extract" Dec 03 17:32:01 crc kubenswrapper[4787]: E1203 17:32:01.939599 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" containerName="pull" Dec 03 17:32:01 crc kubenswrapper[4787]: I1203 17:32:01.939613 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" containerName="pull" Dec 03 17:32:01 crc kubenswrapper[4787]: I1203 17:32:01.939892 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="76ec0f29-6ea6-42a5-a4cc-43d5a8b928be" containerName="extract" Dec 03 17:32:01 crc kubenswrapper[4787]: I1203 17:32:01.940780 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758" Dec 03 17:32:01 crc kubenswrapper[4787]: I1203 17:32:01.942848 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-vbqq9" Dec 03 17:32:01 crc kubenswrapper[4787]: I1203 17:32:01.968963 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758"] Dec 03 17:32:02 crc kubenswrapper[4787]: I1203 17:32:02.102431 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nbzkf\" (UniqueName: \"kubernetes.io/projected/3b816649-4a0d-4090-9e9f-df5c5a296777-kube-api-access-nbzkf\") pod \"openstack-operator-controller-operator-5db84b8f89-tm758\" (UID: \"3b816649-4a0d-4090-9e9f-df5c5a296777\") " pod="openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758" Dec 03 17:32:02 crc kubenswrapper[4787]: I1203 17:32:02.204627 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nbzkf\" (UniqueName: \"kubernetes.io/projected/3b816649-4a0d-4090-9e9f-df5c5a296777-kube-api-access-nbzkf\") pod \"openstack-operator-controller-operator-5db84b8f89-tm758\" (UID: \"3b816649-4a0d-4090-9e9f-df5c5a296777\") " pod="openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758" Dec 03 17:32:02 crc kubenswrapper[4787]: I1203 17:32:02.235721 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nbzkf\" (UniqueName: \"kubernetes.io/projected/3b816649-4a0d-4090-9e9f-df5c5a296777-kube-api-access-nbzkf\") pod \"openstack-operator-controller-operator-5db84b8f89-tm758\" (UID: \"3b816649-4a0d-4090-9e9f-df5c5a296777\") " pod="openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758" Dec 03 17:32:02 crc kubenswrapper[4787]: I1203 17:32:02.265644 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758" Dec 03 17:32:02 crc kubenswrapper[4787]: I1203 17:32:02.752872 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758"] Dec 03 17:32:03 crc kubenswrapper[4787]: I1203 17:32:03.018983 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758" event={"ID":"3b816649-4a0d-4090-9e9f-df5c5a296777","Type":"ContainerStarted","Data":"a8a4a392ab1b05d7908f230265ec5ae3415f73602ab14f3a51009325c1dde7f4"} Dec 03 17:32:09 crc kubenswrapper[4787]: I1203 17:32:09.068834 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758" event={"ID":"3b816649-4a0d-4090-9e9f-df5c5a296777","Type":"ContainerStarted","Data":"6a72ef00e24ed7205ab9a2f2c8d201cac6c7a4298647c35489fb062769b48d59"} Dec 03 17:32:09 crc kubenswrapper[4787]: I1203 17:32:09.069515 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758" Dec 03 17:32:09 crc kubenswrapper[4787]: I1203 17:32:09.105444 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758" podStartSLOduration=2.442514933 podStartE2EDuration="8.105420258s" podCreationTimestamp="2025-12-03 17:32:01 +0000 UTC" firstStartedPulling="2025-12-03 17:32:02.756071698 +0000 UTC m=+1139.573542657" lastFinishedPulling="2025-12-03 17:32:08.418977023 +0000 UTC m=+1145.236447982" observedRunningTime="2025-12-03 17:32:09.099066019 +0000 UTC m=+1145.916536988" watchObservedRunningTime="2025-12-03 17:32:09.105420258 +0000 UTC m=+1145.922891227" Dec 03 17:32:22 crc kubenswrapper[4787]: I1203 17:32:22.272859 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-5db84b8f89-tm758" Dec 03 17:32:40 crc kubenswrapper[4787]: I1203 17:32:40.936152 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn"] Dec 03 17:32:40 crc kubenswrapper[4787]: I1203 17:32:40.938678 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn" Dec 03 17:32:40 crc kubenswrapper[4787]: I1203 17:32:40.941614 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-65mw5" Dec 03 17:32:40 crc kubenswrapper[4787]: I1203 17:32:40.972913 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7"] Dec 03 17:32:40 crc kubenswrapper[4787]: I1203 17:32:40.974301 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7" Dec 03 17:32:40 crc kubenswrapper[4787]: I1203 17:32:40.979142 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-dt5ls" Dec 03 17:32:40 crc kubenswrapper[4787]: I1203 17:32:40.985560 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.017742 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.029603 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.030984 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.036881 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-nx7br" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.049760 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p84h6\" (UniqueName: \"kubernetes.io/projected/6d12a43d-d3da-4b99-b48b-519d660d2527-kube-api-access-p84h6\") pod \"cinder-operator-controller-manager-859b6ccc6-fgdm7\" (UID: \"6d12a43d-d3da-4b99-b48b-519d660d2527\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.049808 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msv9j\" (UniqueName: \"kubernetes.io/projected/43962cfc-342c-49db-83bf-ccde92708a0b-kube-api-access-msv9j\") pod \"barbican-operator-controller-manager-7d9dfd778-bzjxn\" (UID: \"43962cfc-342c-49db-83bf-ccde92708a0b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.074472 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.075822 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.098188 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-b8dp2" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.134777 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.163901 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4fndv\" (UniqueName: \"kubernetes.io/projected/02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23-kube-api-access-4fndv\") pod \"designate-operator-controller-manager-78b4bc895b-mwpr8\" (UID: \"02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.171101 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsbl2\" (UniqueName: \"kubernetes.io/projected/355352cc-c3f1-4430-8ff0-55777890e135-kube-api-access-xsbl2\") pod \"glance-operator-controller-manager-77987cd8cd-p425n\" (UID: \"355352cc-c3f1-4430-8ff0-55777890e135\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.171216 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p84h6\" (UniqueName: \"kubernetes.io/projected/6d12a43d-d3da-4b99-b48b-519d660d2527-kube-api-access-p84h6\") pod \"cinder-operator-controller-manager-859b6ccc6-fgdm7\" (UID: \"6d12a43d-d3da-4b99-b48b-519d660d2527\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.171283 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msv9j\" (UniqueName: \"kubernetes.io/projected/43962cfc-342c-49db-83bf-ccde92708a0b-kube-api-access-msv9j\") pod \"barbican-operator-controller-manager-7d9dfd778-bzjxn\" (UID: \"43962cfc-342c-49db-83bf-ccde92708a0b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.191985 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.200897 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msv9j\" (UniqueName: \"kubernetes.io/projected/43962cfc-342c-49db-83bf-ccde92708a0b-kube-api-access-msv9j\") pod \"barbican-operator-controller-manager-7d9dfd778-bzjxn\" (UID: \"43962cfc-342c-49db-83bf-ccde92708a0b\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.202843 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.204399 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.204906 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p84h6\" (UniqueName: \"kubernetes.io/projected/6d12a43d-d3da-4b99-b48b-519d660d2527-kube-api-access-p84h6\") pod \"cinder-operator-controller-manager-859b6ccc6-fgdm7\" (UID: \"6d12a43d-d3da-4b99-b48b-519d660d2527\") " pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.207978 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-dn5ml" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.236979 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.255193 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-phcn8"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.258780 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.263039 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-rfvjr" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.263191 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.264637 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.267045 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.271807 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-fwjfc" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.273422 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4fndv\" (UniqueName: \"kubernetes.io/projected/02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23-kube-api-access-4fndv\") pod \"designate-operator-controller-manager-78b4bc895b-mwpr8\" (UID: \"02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.273509 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsbl2\" (UniqueName: \"kubernetes.io/projected/355352cc-c3f1-4430-8ff0-55777890e135-kube-api-access-xsbl2\") pod \"glance-operator-controller-manager-77987cd8cd-p425n\" (UID: \"355352cc-c3f1-4430-8ff0-55777890e135\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.273595 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gx485\" (UniqueName: \"kubernetes.io/projected/a6f58c54-821e-4f2b-bce7-1750f64f7ed9-kube-api-access-gx485\") pod \"heat-operator-controller-manager-5f64f6f8bb-hr6rk\" (UID: \"a6f58c54-821e-4f2b-bce7-1750f64f7ed9\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.274889 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.276524 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.298213 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.300652 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-phcn8"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.326264 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsbl2\" (UniqueName: \"kubernetes.io/projected/355352cc-c3f1-4430-8ff0-55777890e135-kube-api-access-xsbl2\") pod \"glance-operator-controller-manager-77987cd8cd-p425n\" (UID: \"355352cc-c3f1-4430-8ff0-55777890e135\") " pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.333513 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4fndv\" (UniqueName: \"kubernetes.io/projected/02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23-kube-api-access-4fndv\") pod \"designate-operator-controller-manager-78b4bc895b-mwpr8\" (UID: \"02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23\") " pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.354567 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.360065 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.361606 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.376725 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-pfbf7" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.377828 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9th54\" (UniqueName: \"kubernetes.io/projected/2a789aa2-cc4a-42a9-b1c0-9d1a648fd509-kube-api-access-9th54\") pod \"horizon-operator-controller-manager-68c6d99b8f-cjbzf\" (UID: \"2a789aa2-cc4a-42a9-b1c0-9d1a648fd509\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.377928 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.378080 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gx485\" (UniqueName: \"kubernetes.io/projected/a6f58c54-821e-4f2b-bce7-1750f64f7ed9-kube-api-access-gx485\") pod \"heat-operator-controller-manager-5f64f6f8bb-hr6rk\" (UID: \"a6f58c54-821e-4f2b-bce7-1750f64f7ed9\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.378202 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgfwr\" (UniqueName: \"kubernetes.io/projected/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-kube-api-access-pgfwr\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.412716 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.414866 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.429543 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-vhlfb" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.439569 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.443733 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.466711 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gx485\" (UniqueName: \"kubernetes.io/projected/a6f58c54-821e-4f2b-bce7-1750f64f7ed9-kube-api-access-gx485\") pod \"heat-operator-controller-manager-5f64f6f8bb-hr6rk\" (UID: \"a6f58c54-821e-4f2b-bce7-1750f64f7ed9\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.469987 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.480690 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgfwr\" (UniqueName: \"kubernetes.io/projected/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-kube-api-access-pgfwr\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.480824 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9th54\" (UniqueName: \"kubernetes.io/projected/2a789aa2-cc4a-42a9-b1c0-9d1a648fd509-kube-api-access-9th54\") pod \"horizon-operator-controller-manager-68c6d99b8f-cjbzf\" (UID: \"2a789aa2-cc4a-42a9-b1c0-9d1a648fd509\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.480857 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcvms\" (UniqueName: \"kubernetes.io/projected/f94890f3-3499-483d-9cda-24a377e982a8-kube-api-access-gcvms\") pod \"keystone-operator-controller-manager-7765d96ddf-6txhw\" (UID: \"f94890f3-3499-483d-9cda-24a377e982a8\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.480906 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.480957 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzgpd\" (UniqueName: \"kubernetes.io/projected/8c4a1981-21a3-49ec-a82d-419f8c080fde-kube-api-access-mzgpd\") pod \"ironic-operator-controller-manager-6c548fd776-bgmsg\" (UID: \"8c4a1981-21a3-49ec-a82d-419f8c080fde\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.481490 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" Dec 03 17:32:41 crc kubenswrapper[4787]: E1203 17:32:41.482206 4787 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:41 crc kubenswrapper[4787]: E1203 17:32:41.482252 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert podName:b514a6d2-89f9-451c-af24-c0c5a49cdd8d nodeName:}" failed. No retries permitted until 2025-12-03 17:32:41.982236879 +0000 UTC m=+1178.799707838 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert") pod "infra-operator-controller-manager-57548d458d-phcn8" (UID: "b514a6d2-89f9-451c-af24-c0c5a49cdd8d") : secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.487109 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-l946f" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.521637 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.525436 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9th54\" (UniqueName: \"kubernetes.io/projected/2a789aa2-cc4a-42a9-b1c0-9d1a648fd509-kube-api-access-9th54\") pod \"horizon-operator-controller-manager-68c6d99b8f-cjbzf\" (UID: \"2a789aa2-cc4a-42a9-b1c0-9d1a648fd509\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.526904 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgfwr\" (UniqueName: \"kubernetes.io/projected/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-kube-api-access-pgfwr\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.539970 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.558538 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.565885 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.572777 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.574089 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-cm52n" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.583833 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l5p9\" (UniqueName: \"kubernetes.io/projected/5c7f6976-b7e2-453d-ad11-6d0e978cd03d-kube-api-access-6l5p9\") pod \"manila-operator-controller-manager-7c79b5df47-cd9jr\" (UID: \"5c7f6976-b7e2-453d-ad11-6d0e978cd03d\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.583967 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcvms\" (UniqueName: \"kubernetes.io/projected/f94890f3-3499-483d-9cda-24a377e982a8-kube-api-access-gcvms\") pod \"keystone-operator-controller-manager-7765d96ddf-6txhw\" (UID: \"f94890f3-3499-483d-9cda-24a377e982a8\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.584109 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzgpd\" (UniqueName: \"kubernetes.io/projected/8c4a1981-21a3-49ec-a82d-419f8c080fde-kube-api-access-mzgpd\") pod \"ironic-operator-controller-manager-6c548fd776-bgmsg\" (UID: \"8c4a1981-21a3-49ec-a82d-419f8c080fde\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.604586 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.606092 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.617419 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-p6wxd" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.627718 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzgpd\" (UniqueName: \"kubernetes.io/projected/8c4a1981-21a3-49ec-a82d-419f8c080fde-kube-api-access-mzgpd\") pod \"ironic-operator-controller-manager-6c548fd776-bgmsg\" (UID: \"8c4a1981-21a3-49ec-a82d-419f8c080fde\") " pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.635312 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.636873 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.639530 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcvms\" (UniqueName: \"kubernetes.io/projected/f94890f3-3499-483d-9cda-24a377e982a8-kube-api-access-gcvms\") pod \"keystone-operator-controller-manager-7765d96ddf-6txhw\" (UID: \"f94890f3-3499-483d-9cda-24a377e982a8\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.647905 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-mcn7c" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.685898 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5ljcj\" (UniqueName: \"kubernetes.io/projected/04fbeb64-47f2-4393-8295-0ac7d6db4df3-kube-api-access-5ljcj\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-sbcbk\" (UID: \"04fbeb64-47f2-4393-8295-0ac7d6db4df3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.686143 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52vmt\" (UniqueName: \"kubernetes.io/projected/a0326a22-7690-48ed-84f5-01cc9e8331b2-kube-api-access-52vmt\") pod \"nova-operator-controller-manager-697bc559fc-zwmxn\" (UID: \"a0326a22-7690-48ed-84f5-01cc9e8331b2\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.686171 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wldnb\" (UniqueName: \"kubernetes.io/projected/6384c305-657a-465d-8e48-487fb12cd52f-kube-api-access-wldnb\") pod \"mariadb-operator-controller-manager-56bbcc9d85-5m5h7\" (UID: \"6384c305-657a-465d-8e48-487fb12cd52f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.686225 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l5p9\" (UniqueName: \"kubernetes.io/projected/5c7f6976-b7e2-453d-ad11-6d0e978cd03d-kube-api-access-6l5p9\") pod \"manila-operator-controller-manager-7c79b5df47-cd9jr\" (UID: \"5c7f6976-b7e2-453d-ad11-6d0e978cd03d\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.694309 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.709902 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.715781 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-75r2j"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.717106 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-75r2j" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.727387 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-5k4rn" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.750514 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.769065 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-75r2j"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.776771 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l5p9\" (UniqueName: \"kubernetes.io/projected/5c7f6976-b7e2-453d-ad11-6d0e978cd03d-kube-api-access-6l5p9\") pod \"manila-operator-controller-manager-7c79b5df47-cd9jr\" (UID: \"5c7f6976-b7e2-453d-ad11-6d0e978cd03d\") " pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.788750 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vlsm\" (UniqueName: \"kubernetes.io/projected/67472bf9-12b8-4463-bc95-dec19e689f36-kube-api-access-4vlsm\") pod \"octavia-operator-controller-manager-998648c74-75r2j\" (UID: \"67472bf9-12b8-4463-bc95-dec19e689f36\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-75r2j" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.788851 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5ljcj\" (UniqueName: \"kubernetes.io/projected/04fbeb64-47f2-4393-8295-0ac7d6db4df3-kube-api-access-5ljcj\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-sbcbk\" (UID: \"04fbeb64-47f2-4393-8295-0ac7d6db4df3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.788882 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52vmt\" (UniqueName: \"kubernetes.io/projected/a0326a22-7690-48ed-84f5-01cc9e8331b2-kube-api-access-52vmt\") pod \"nova-operator-controller-manager-697bc559fc-zwmxn\" (UID: \"a0326a22-7690-48ed-84f5-01cc9e8331b2\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.789174 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wldnb\" (UniqueName: \"kubernetes.io/projected/6384c305-657a-465d-8e48-487fb12cd52f-kube-api-access-wldnb\") pod \"mariadb-operator-controller-manager-56bbcc9d85-5m5h7\" (UID: \"6384c305-657a-465d-8e48-487fb12cd52f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.819049 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52vmt\" (UniqueName: \"kubernetes.io/projected/a0326a22-7690-48ed-84f5-01cc9e8331b2-kube-api-access-52vmt\") pod \"nova-operator-controller-manager-697bc559fc-zwmxn\" (UID: \"a0326a22-7690-48ed-84f5-01cc9e8331b2\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.821571 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5ljcj\" (UniqueName: \"kubernetes.io/projected/04fbeb64-47f2-4393-8295-0ac7d6db4df3-kube-api-access-5ljcj\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-sbcbk\" (UID: \"04fbeb64-47f2-4393-8295-0ac7d6db4df3\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.832757 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.854917 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wldnb\" (UniqueName: \"kubernetes.io/projected/6384c305-657a-465d-8e48-487fb12cd52f-kube-api-access-wldnb\") pod \"mariadb-operator-controller-manager-56bbcc9d85-5m5h7\" (UID: \"6384c305-657a-465d-8e48-487fb12cd52f\") " pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.858397 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.874492 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.890787 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vlsm\" (UniqueName: \"kubernetes.io/projected/67472bf9-12b8-4463-bc95-dec19e689f36-kube-api-access-4vlsm\") pod \"octavia-operator-controller-manager-998648c74-75r2j\" (UID: \"67472bf9-12b8-4463-bc95-dec19e689f36\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-75r2j" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.922416 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.925528 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.938673 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.938903 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.947134 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-7l5tk" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.948125 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.956069 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz"] Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.956289 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.957374 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.961101 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-q67dx" Dec 03 17:32:41 crc kubenswrapper[4787]: I1203 17:32:41.992305 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vlsm\" (UniqueName: \"kubernetes.io/projected/67472bf9-12b8-4463-bc95-dec19e689f36-kube-api-access-4vlsm\") pod \"octavia-operator-controller-manager-998648c74-75r2j\" (UID: \"67472bf9-12b8-4463-bc95-dec19e689f36\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-75r2j" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.001302 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wcw2m\" (UniqueName: \"kubernetes.io/projected/31c5513f-61f7-4d6c-8136-b212cc171732-kube-api-access-wcw2m\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.001409 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4tsgb\" (UniqueName: \"kubernetes.io/projected/226cdcb0-0abc-4a06-9d5c-1d30d9ac8288-kube-api-access-4tsgb\") pod \"ovn-operator-controller-manager-b6456fdb6-xg7kr\" (UID: \"226cdcb0-0abc-4a06-9d5c-1d30d9ac8288\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.001458 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.001489 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:32:42 crc kubenswrapper[4787]: E1203 17:32:42.002076 4787 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:42 crc kubenswrapper[4787]: E1203 17:32:42.002228 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert podName:b514a6d2-89f9-451c-af24-c0c5a49cdd8d nodeName:}" failed. No retries permitted until 2025-12-03 17:32:43.002191881 +0000 UTC m=+1179.819663000 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert") pod "infra-operator-controller-manager-57548d458d-phcn8" (UID: "b514a6d2-89f9-451c-af24-c0c5a49cdd8d") : secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.019540 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.022270 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.027362 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-wv58t" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.031353 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.033820 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.042517 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.048842 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-2xm6q" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.050308 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.074105 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.078879 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.094186 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.102974 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wcw2m\" (UniqueName: \"kubernetes.io/projected/31c5513f-61f7-4d6c-8136-b212cc171732-kube-api-access-wcw2m\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.103148 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wz8pd\" (UniqueName: \"kubernetes.io/projected/dd3a3971-adb9-46cb-a0b7-63897e326341-kube-api-access-wz8pd\") pod \"placement-operator-controller-manager-78f8948974-zgkn8\" (UID: \"dd3a3971-adb9-46cb-a0b7-63897e326341\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.103183 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4tsgb\" (UniqueName: \"kubernetes.io/projected/226cdcb0-0abc-4a06-9d5c-1d30d9ac8288-kube-api-access-4tsgb\") pod \"ovn-operator-controller-manager-b6456fdb6-xg7kr\" (UID: \"226cdcb0-0abc-4a06-9d5c-1d30d9ac8288\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.103242 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.103264 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-klqkr\" (UniqueName: \"kubernetes.io/projected/c209ad3f-9014-4385-96cb-323dd27b8335-kube-api-access-klqkr\") pod \"swift-operator-controller-manager-5f8c65bbfc-vmpfz\" (UID: \"c209ad3f-9014-4385-96cb-323dd27b8335\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" Dec 03 17:32:42 crc kubenswrapper[4787]: E1203 17:32:42.104582 4787 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:42 crc kubenswrapper[4787]: E1203 17:32:42.104644 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert podName:31c5513f-61f7-4d6c-8136-b212cc171732 nodeName:}" failed. No retries permitted until 2025-12-03 17:32:42.60462102 +0000 UTC m=+1179.422092199 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" (UID: "31c5513f-61f7-4d6c-8136-b212cc171732") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.114050 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.115362 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.120420 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-75r2j" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.121103 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-jcjvf" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.146308 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.153187 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4tsgb\" (UniqueName: \"kubernetes.io/projected/226cdcb0-0abc-4a06-9d5c-1d30d9ac8288-kube-api-access-4tsgb\") pod \"ovn-operator-controller-manager-b6456fdb6-xg7kr\" (UID: \"226cdcb0-0abc-4a06-9d5c-1d30d9ac8288\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.163534 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wcw2m\" (UniqueName: \"kubernetes.io/projected/31c5513f-61f7-4d6c-8136-b212cc171732-kube-api-access-wcw2m\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.192561 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.200614 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.202451 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.206045 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-kjqk6" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.208899 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wz8pd\" (UniqueName: \"kubernetes.io/projected/dd3a3971-adb9-46cb-a0b7-63897e326341-kube-api-access-wz8pd\") pod \"placement-operator-controller-manager-78f8948974-zgkn8\" (UID: \"dd3a3971-adb9-46cb-a0b7-63897e326341\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.209099 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-klqkr\" (UniqueName: \"kubernetes.io/projected/c209ad3f-9014-4385-96cb-323dd27b8335-kube-api-access-klqkr\") pod \"swift-operator-controller-manager-5f8c65bbfc-vmpfz\" (UID: \"c209ad3f-9014-4385-96cb-323dd27b8335\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.209177 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6mcd9\" (UniqueName: \"kubernetes.io/projected/a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173-kube-api-access-6mcd9\") pod \"telemetry-operator-controller-manager-66f6f55998-k4kg2\" (UID: \"a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173\") " pod="openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.224459 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.266298 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.268235 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.274940 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-qcnsp" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.290838 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-klqkr\" (UniqueName: \"kubernetes.io/projected/c209ad3f-9014-4385-96cb-323dd27b8335-kube-api-access-klqkr\") pod \"swift-operator-controller-manager-5f8c65bbfc-vmpfz\" (UID: \"c209ad3f-9014-4385-96cb-323dd27b8335\") " pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.303675 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.299149 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wz8pd\" (UniqueName: \"kubernetes.io/projected/dd3a3971-adb9-46cb-a0b7-63897e326341-kube-api-access-wz8pd\") pod \"placement-operator-controller-manager-78f8948974-zgkn8\" (UID: \"dd3a3971-adb9-46cb-a0b7-63897e326341\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.327169 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6mcd9\" (UniqueName: \"kubernetes.io/projected/a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173-kube-api-access-6mcd9\") pod \"telemetry-operator-controller-manager-66f6f55998-k4kg2\" (UID: \"a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173\") " pod="openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.327293 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8ljjw\" (UniqueName: \"kubernetes.io/projected/d1364502-c591-47c8-b24c-00c77e3a8d23-kube-api-access-8ljjw\") pod \"test-operator-controller-manager-5854674fcc-wtvd4\" (UID: \"d1364502-c591-47c8-b24c-00c77e3a8d23\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.327334 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tnmlq\" (UniqueName: \"kubernetes.io/projected/8af08022-68dd-4fc4-a052-2cfde5f1ab0a-kube-api-access-tnmlq\") pod \"watcher-operator-controller-manager-769dc69bc-2wq6s\" (UID: \"8af08022-68dd-4fc4-a052-2cfde5f1ab0a\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.341803 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.347595 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.354805 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.364433 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6mcd9\" (UniqueName: \"kubernetes.io/projected/a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173-kube-api-access-6mcd9\") pod \"telemetry-operator-controller-manager-66f6f55998-k4kg2\" (UID: \"a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173\") " pod="openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.379806 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-tg82f" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.380448 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.380528 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.428329 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.430702 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.431950 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hspgq\" (UniqueName: \"kubernetes.io/projected/d9b27877-9ba0-4487-a487-aceafca2075a-kube-api-access-hspgq\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.432049 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8ljjw\" (UniqueName: \"kubernetes.io/projected/d1364502-c591-47c8-b24c-00c77e3a8d23-kube-api-access-8ljjw\") pod \"test-operator-controller-manager-5854674fcc-wtvd4\" (UID: \"d1364502-c591-47c8-b24c-00c77e3a8d23\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.432077 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tnmlq\" (UniqueName: \"kubernetes.io/projected/8af08022-68dd-4fc4-a052-2cfde5f1ab0a-kube-api-access-tnmlq\") pod \"watcher-operator-controller-manager-769dc69bc-2wq6s\" (UID: \"8af08022-68dd-4fc4-a052-2cfde5f1ab0a\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.432155 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.432180 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.436417 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.438273 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-q6hmw" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.443930 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.445993 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.462802 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8ljjw\" (UniqueName: \"kubernetes.io/projected/d1364502-c591-47c8-b24c-00c77e3a8d23-kube-api-access-8ljjw\") pod \"test-operator-controller-manager-5854674fcc-wtvd4\" (UID: \"d1364502-c591-47c8-b24c-00c77e3a8d23\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.467414 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.467618 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tnmlq\" (UniqueName: \"kubernetes.io/projected/8af08022-68dd-4fc4-a052-2cfde5f1ab0a-kube-api-access-tnmlq\") pod \"watcher-operator-controller-manager-769dc69bc-2wq6s\" (UID: \"8af08022-68dd-4fc4-a052-2cfde5f1ab0a\") " pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.500709 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.533444 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.533811 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.533860 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hspgq\" (UniqueName: \"kubernetes.io/projected/d9b27877-9ba0-4487-a487-aceafca2075a-kube-api-access-hspgq\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.533939 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsndp\" (UniqueName: \"kubernetes.io/projected/c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1-kube-api-access-hsndp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cnjbp\" (UID: \"c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp" Dec 03 17:32:42 crc kubenswrapper[4787]: E1203 17:32:42.533737 4787 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 17:32:42 crc kubenswrapper[4787]: E1203 17:32:42.534167 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:32:43.034147631 +0000 UTC m=+1179.851618590 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "metrics-server-cert" not found Dec 03 17:32:42 crc kubenswrapper[4787]: E1203 17:32:42.534099 4787 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:32:42 crc kubenswrapper[4787]: E1203 17:32:42.534434 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:32:43.034416158 +0000 UTC m=+1179.851887117 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "webhook-server-cert" not found Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.558284 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.559908 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.562597 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hspgq\" (UniqueName: \"kubernetes.io/projected/d9b27877-9ba0-4487-a487-aceafca2075a-kube-api-access-hspgq\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.636450 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.636611 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsndp\" (UniqueName: \"kubernetes.io/projected/c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1-kube-api-access-hsndp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cnjbp\" (UID: \"c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp" Dec 03 17:32:42 crc kubenswrapper[4787]: E1203 17:32:42.637301 4787 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:42 crc kubenswrapper[4787]: E1203 17:32:42.637353 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert podName:31c5513f-61f7-4d6c-8136-b212cc171732 nodeName:}" failed. No retries permitted until 2025-12-03 17:32:43.63733424 +0000 UTC m=+1180.454805199 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" (UID: "31c5513f-61f7-4d6c-8136-b212cc171732") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.684939 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsndp\" (UniqueName: \"kubernetes.io/projected/c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1-kube-api-access-hsndp\") pod \"rabbitmq-cluster-operator-manager-668c99d594-cnjbp\" (UID: \"c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.785221 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.798752 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n"] Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.902204 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp" Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.983555 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7"] Dec 03 17:32:42 crc kubenswrapper[4787]: W1203 17:32:42.987211 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6d12a43d_d3da_4b99_b48b_519d660d2527.slice/crio-452353cfb338cc3576d3d177097392eb98f7974aada8118af97cacf0ac5d7765 WatchSource:0}: Error finding container 452353cfb338cc3576d3d177097392eb98f7974aada8118af97cacf0ac5d7765: Status 404 returned error can't find the container with id 452353cfb338cc3576d3d177097392eb98f7974aada8118af97cacf0ac5d7765 Dec 03 17:32:42 crc kubenswrapper[4787]: W1203 17:32:42.994077 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6f58c54_821e_4f2b_bce7_1750f64f7ed9.slice/crio-5fd4750976aced0c509c38e269cc5dfaac6b882df8970489ab6909c9aaabb7ee WatchSource:0}: Error finding container 5fd4750976aced0c509c38e269cc5dfaac6b882df8970489ab6909c9aaabb7ee: Status 404 returned error can't find the container with id 5fd4750976aced0c509c38e269cc5dfaac6b882df8970489ab6909c9aaabb7ee Dec 03 17:32:42 crc kubenswrapper[4787]: I1203 17:32:42.995134 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.008279 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.050709 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.050795 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.050840 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.050910 4787 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.050995 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert podName:b514a6d2-89f9-451c-af24-c0c5a49cdd8d nodeName:}" failed. No retries permitted until 2025-12-03 17:32:45.050971249 +0000 UTC m=+1181.868442248 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert") pod "infra-operator-controller-manager-57548d458d-phcn8" (UID: "b514a6d2-89f9-451c-af24-c0c5a49cdd8d") : secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.051154 4787 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.051201 4787 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.051269 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:32:44.051250056 +0000 UTC m=+1180.868721095 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "webhook-server-cert" not found Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.051295 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:32:44.051282847 +0000 UTC m=+1180.868753916 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "metrics-server-cert" not found Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.427124 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.431859 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" event={"ID":"02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23","Type":"ContainerStarted","Data":"8e6ac2c765e763a5af2c845debf20bfcedcf79f3a8dd4b692bb9aa6d78fdea43"} Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.450429 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn" event={"ID":"43962cfc-342c-49db-83bf-ccde92708a0b","Type":"ContainerStarted","Data":"fcbdf53ad97b706a686607a81aa6125bd68b83c774d1907a1ac5c7161ad0c766"} Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.451346 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg"] Dec 03 17:32:43 crc kubenswrapper[4787]: W1203 17:32:43.452096 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod226cdcb0_0abc_4a06_9d5c_1d30d9ac8288.slice/crio-244b1a5ac745d52cd50161732227864c5c499ad201552d0da9ce19660e204487 WatchSource:0}: Error finding container 244b1a5ac745d52cd50161732227864c5c499ad201552d0da9ce19660e204487: Status 404 returned error can't find the container with id 244b1a5ac745d52cd50161732227864c5c499ad201552d0da9ce19660e204487 Dec 03 17:32:43 crc kubenswrapper[4787]: W1203 17:32:43.452305 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6384c305_657a_465d_8e48_487fb12cd52f.slice/crio-f77af1ca3e166ff820eb26c7f59cbbc37d796b60547262a904c4387b5d41c8e9 WatchSource:0}: Error finding container f77af1ca3e166ff820eb26c7f59cbbc37d796b60547262a904c4387b5d41c8e9: Status 404 returned error can't find the container with id f77af1ca3e166ff820eb26c7f59cbbc37d796b60547262a904c4387b5d41c8e9 Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.461493 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.462666 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk" event={"ID":"a6f58c54-821e-4f2b-bce7-1750f64f7ed9","Type":"ContainerStarted","Data":"5fd4750976aced0c509c38e269cc5dfaac6b882df8970489ab6909c9aaabb7ee"} Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.478565 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.486793 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.486843 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" event={"ID":"a0326a22-7690-48ed-84f5-01cc9e8331b2","Type":"ContainerStarted","Data":"c37e82bc1fe9f354dcc875e16876f8a093065445ad1b76c85ececd746470cbe0"} Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.493902 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7" event={"ID":"6d12a43d-d3da-4b99-b48b-519d660d2527","Type":"ContainerStarted","Data":"452353cfb338cc3576d3d177097392eb98f7974aada8118af97cacf0ac5d7765"} Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.495241 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n" event={"ID":"355352cc-c3f1-4430-8ff0-55777890e135","Type":"ContainerStarted","Data":"dfc23a7bd8fad0f1cce2853ec4c77f942f6dbe32c06f929d05f6c29e60a0def1"} Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.499821 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf"] Dec 03 17:32:43 crc kubenswrapper[4787]: W1203 17:32:43.551171 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2a789aa2_cc4a_42a9_b1c0_9d1a648fd509.slice/crio-bc24b2073ce58863211ee5e362716172c59e0564d938b07dae8278394550a713 WatchSource:0}: Error finding container bc24b2073ce58863211ee5e362716172c59e0564d938b07dae8278394550a713: Status 404 returned error can't find the container with id bc24b2073ce58863211ee5e362716172c59e0564d938b07dae8278394550a713 Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.667305 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.670152 4787 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.670922 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert podName:31c5513f-61f7-4d6c-8136-b212cc171732 nodeName:}" failed. No retries permitted until 2025-12-03 17:32:45.670902435 +0000 UTC m=+1182.488373394 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" (UID: "31c5513f-61f7-4d6c-8136-b212cc171732") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.681211 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.689221 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.726110 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.739118 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-75r2j"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.747142 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.756556 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp"] Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.762427 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4"] Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.795630 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tnmlq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-2wq6s_openstack-operators(8af08022-68dd-4fc4-a052-2cfde5f1ab0a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.799147 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8ljjw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-wtvd4_openstack-operators(d1364502-c591-47c8-b24c-00c77e3a8d23): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.804571 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tnmlq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-769dc69bc-2wq6s_openstack-operators(8af08022-68dd-4fc4-a052-2cfde5f1ab0a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.804677 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-klqkr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-vmpfz_openstack-operators(c209ad3f-9014-4385-96cb-323dd27b8335): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.804795 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6l5p9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-cd9jr_openstack-operators(5c7f6976-b7e2-453d-ad11-6d0e978cd03d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.804874 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8ljjw,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-wtvd4_openstack-operators(d1364502-c591-47c8-b24c-00c77e3a8d23): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.806326 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" podUID="d1364502-c591-47c8-b24c-00c77e3a8d23" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.806507 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" podUID="8af08022-68dd-4fc4-a052-2cfde5f1ab0a" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.809299 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6l5p9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-7c79b5df47-cd9jr_openstack-operators(5c7f6976-b7e2-453d-ad11-6d0e978cd03d): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.815866 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-klqkr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-vmpfz_openstack-operators(c209ad3f-9014-4385-96cb-323dd27b8335): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.814766 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" podUID="5c7f6976-b7e2-453d-ad11-6d0e978cd03d" Dec 03 17:32:43 crc kubenswrapper[4787]: E1203 17:32:43.819850 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" podUID="c209ad3f-9014-4385-96cb-323dd27b8335" Dec 03 17:32:43 crc kubenswrapper[4787]: I1203 17:32:43.823369 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s"] Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.088789 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.088886 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:44 crc kubenswrapper[4787]: E1203 17:32:44.089293 4787 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 17:32:44 crc kubenswrapper[4787]: E1203 17:32:44.089330 4787 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:32:44 crc kubenswrapper[4787]: E1203 17:32:44.089389 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:32:46.089365533 +0000 UTC m=+1182.906836682 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "metrics-server-cert" not found Dec 03 17:32:44 crc kubenswrapper[4787]: E1203 17:32:44.089414 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:32:46.089404854 +0000 UTC m=+1182.906876023 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "webhook-server-cert" not found Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.529327 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg" event={"ID":"8c4a1981-21a3-49ec-a82d-419f8c080fde","Type":"ContainerStarted","Data":"154eb6561d06409a6d34980658e50bf5a92e2d576cc7572fa0585e8ba8d13437"} Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.535742 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2" event={"ID":"a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173","Type":"ContainerStarted","Data":"692e3fb8b51b60ee85f3f688be2a07f54004d2941e8d84cbbb8714b2286b12f2"} Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.540309 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp" event={"ID":"c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1","Type":"ContainerStarted","Data":"b8e95739ccae667734df72bb47baecb3f59a882726575db18986796a604d355d"} Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.543824 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" event={"ID":"f94890f3-3499-483d-9cda-24a377e982a8","Type":"ContainerStarted","Data":"daed7e5a7f6a348624759fa2dd7134240e3b7c1dc159b33f9eae6dc990830752"} Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.545614 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" event={"ID":"6384c305-657a-465d-8e48-487fb12cd52f","Type":"ContainerStarted","Data":"f77af1ca3e166ff820eb26c7f59cbbc37d796b60547262a904c4387b5d41c8e9"} Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.547987 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" event={"ID":"d1364502-c591-47c8-b24c-00c77e3a8d23","Type":"ContainerStarted","Data":"700c38c0565f27f3b9d909aaf671572c63d875836291d990fbc8f42f458a7429"} Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.552050 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" event={"ID":"2a789aa2-cc4a-42a9-b1c0-9d1a648fd509","Type":"ContainerStarted","Data":"bc24b2073ce58863211ee5e362716172c59e0564d938b07dae8278394550a713"} Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.554532 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" event={"ID":"8af08022-68dd-4fc4-a052-2cfde5f1ab0a","Type":"ContainerStarted","Data":"be568dad6fdac55fb73418686ec94f06e886876b9bf9a82e5ca8dcb53dd5a314"} Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.560034 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" event={"ID":"04fbeb64-47f2-4393-8295-0ac7d6db4df3","Type":"ContainerStarted","Data":"b44c73776253846a1c90d84bc3ddb847e23c283f7a7c85da471b0918e7747015"} Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.562400 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-75r2j" event={"ID":"67472bf9-12b8-4463-bc95-dec19e689f36","Type":"ContainerStarted","Data":"eee66e43242c40307ddc9957705a17eeaf4745e0994c62e76509da02179b5a89"} Dec 03 17:32:44 crc kubenswrapper[4787]: E1203 17:32:44.575796 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" podUID="d1364502-c591-47c8-b24c-00c77e3a8d23" Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.576925 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr" event={"ID":"226cdcb0-0abc-4a06-9d5c-1d30d9ac8288","Type":"ContainerStarted","Data":"244b1a5ac745d52cd50161732227864c5c499ad201552d0da9ce19660e204487"} Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.583206 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" event={"ID":"c209ad3f-9014-4385-96cb-323dd27b8335","Type":"ContainerStarted","Data":"99609a73e77f13b882a2df43d9b2b7017713cbbe3bf63f86e3c71361afea70ab"} Dec 03 17:32:44 crc kubenswrapper[4787]: E1203 17:32:44.593768 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" podUID="c209ad3f-9014-4385-96cb-323dd27b8335" Dec 03 17:32:44 crc kubenswrapper[4787]: E1203 17:32:44.594074 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" podUID="8af08022-68dd-4fc4-a052-2cfde5f1ab0a" Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.600788 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" event={"ID":"5c7f6976-b7e2-453d-ad11-6d0e978cd03d","Type":"ContainerStarted","Data":"8698059392d2c5e642945b48a5b4d48a87a79d4e8e16b26f2fa26c209c20d056"} Dec 03 17:32:44 crc kubenswrapper[4787]: E1203 17:32:44.607754 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" podUID="5c7f6976-b7e2-453d-ad11-6d0e978cd03d" Dec 03 17:32:44 crc kubenswrapper[4787]: I1203 17:32:44.608832 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8" event={"ID":"dd3a3971-adb9-46cb-a0b7-63897e326341","Type":"ContainerStarted","Data":"308e8a1a359892acee9492d20932f787b7420559b91fcece75d1aa19fb417bc9"} Dec 03 17:32:45 crc kubenswrapper[4787]: I1203 17:32:45.121171 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:32:45 crc kubenswrapper[4787]: E1203 17:32:45.121381 4787 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:45 crc kubenswrapper[4787]: E1203 17:32:45.121959 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert podName:b514a6d2-89f9-451c-af24-c0c5a49cdd8d nodeName:}" failed. No retries permitted until 2025-12-03 17:32:49.121920442 +0000 UTC m=+1185.939391501 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert") pod "infra-operator-controller-manager-57548d458d-phcn8" (UID: "b514a6d2-89f9-451c-af24-c0c5a49cdd8d") : secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:45 crc kubenswrapper[4787]: E1203 17:32:45.651744 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" podUID="d1364502-c591-47c8-b24c-00c77e3a8d23" Dec 03 17:32:45 crc kubenswrapper[4787]: E1203 17:32:45.651848 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:2e59cfbeefc3aff0bb0a6ae9ce2235129f5173c98dd5ee8dac229ad4895faea9\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" podUID="5c7f6976-b7e2-453d-ad11-6d0e978cd03d" Dec 03 17:32:45 crc kubenswrapper[4787]: E1203 17:32:45.651925 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:9aa8c03633e4b934c57868c1660acf47e7d386ac86bcb344df262c9ad76b8621\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" podUID="8af08022-68dd-4fc4-a052-2cfde5f1ab0a" Dec 03 17:32:45 crc kubenswrapper[4787]: E1203 17:32:45.651984 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:2a3d21728a8bfb4e64617e63e61e2d1cb70a383ea3e8f846e0c3c3c02d2b0a9d\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" podUID="c209ad3f-9014-4385-96cb-323dd27b8335" Dec 03 17:32:45 crc kubenswrapper[4787]: I1203 17:32:45.740027 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:32:45 crc kubenswrapper[4787]: E1203 17:32:45.740279 4787 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:45 crc kubenswrapper[4787]: E1203 17:32:45.740391 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert podName:31c5513f-61f7-4d6c-8136-b212cc171732 nodeName:}" failed. No retries permitted until 2025-12-03 17:32:49.740369129 +0000 UTC m=+1186.557840088 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" (UID: "31c5513f-61f7-4d6c-8136-b212cc171732") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:46 crc kubenswrapper[4787]: I1203 17:32:46.146616 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:46 crc kubenswrapper[4787]: I1203 17:32:46.146918 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:46 crc kubenswrapper[4787]: E1203 17:32:46.146783 4787 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 17:32:46 crc kubenswrapper[4787]: E1203 17:32:46.147208 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:32:50.147193779 +0000 UTC m=+1186.964664738 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "metrics-server-cert" not found Dec 03 17:32:46 crc kubenswrapper[4787]: E1203 17:32:46.147152 4787 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:32:46 crc kubenswrapper[4787]: E1203 17:32:46.148351 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:32:50.148313319 +0000 UTC m=+1186.965784458 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "webhook-server-cert" not found Dec 03 17:32:49 crc kubenswrapper[4787]: I1203 17:32:49.203292 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:32:49 crc kubenswrapper[4787]: E1203 17:32:49.203452 4787 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:49 crc kubenswrapper[4787]: E1203 17:32:49.203808 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert podName:b514a6d2-89f9-451c-af24-c0c5a49cdd8d nodeName:}" failed. No retries permitted until 2025-12-03 17:32:57.20378604 +0000 UTC m=+1194.021257009 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert") pod "infra-operator-controller-manager-57548d458d-phcn8" (UID: "b514a6d2-89f9-451c-af24-c0c5a49cdd8d") : secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:49 crc kubenswrapper[4787]: I1203 17:32:49.814639 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:32:49 crc kubenswrapper[4787]: E1203 17:32:49.814850 4787 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:49 crc kubenswrapper[4787]: E1203 17:32:49.814976 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert podName:31c5513f-61f7-4d6c-8136-b212cc171732 nodeName:}" failed. No retries permitted until 2025-12-03 17:32:57.814948135 +0000 UTC m=+1194.632419154 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" (UID: "31c5513f-61f7-4d6c-8136-b212cc171732") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:50 crc kubenswrapper[4787]: I1203 17:32:50.230899 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:50 crc kubenswrapper[4787]: I1203 17:32:50.231343 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:50 crc kubenswrapper[4787]: E1203 17:32:50.231215 4787 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 17:32:50 crc kubenswrapper[4787]: E1203 17:32:50.231479 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:32:58.231436959 +0000 UTC m=+1195.048907918 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "metrics-server-cert" not found Dec 03 17:32:50 crc kubenswrapper[4787]: E1203 17:32:50.231524 4787 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:32:50 crc kubenswrapper[4787]: E1203 17:32:50.231641 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:32:58.231608734 +0000 UTC m=+1195.049079683 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "webhook-server-cert" not found Dec 03 17:32:56 crc kubenswrapper[4787]: E1203 17:32:56.143986 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85" Dec 03 17:32:56 crc kubenswrapper[4787]: E1203 17:32:56.144769 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/designate-operator@sha256:9f68d7bc8c6bce38f46dee8a8272d5365c49fe7b32b2af52e8ac884e212f3a85,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-4fndv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod designate-operator-controller-manager-78b4bc895b-mwpr8_openstack-operators(02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:32:56 crc kubenswrapper[4787]: E1203 17:32:56.793151 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7" Dec 03 17:32:56 crc kubenswrapper[4787]: E1203 17:32:56.793588 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:600ca007e493d3af0fcc2ebac92e8da5efd2afe812b62d7d3d4dd0115bdf05d7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wldnb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-5m5h7_openstack-operators(6384c305-657a-465d-8e48-487fb12cd52f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:32:57 crc kubenswrapper[4787]: I1203 17:32:57.267653 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:32:57 crc kubenswrapper[4787]: E1203 17:32:57.267851 4787 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:57 crc kubenswrapper[4787]: E1203 17:32:57.268419 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert podName:b514a6d2-89f9-451c-af24-c0c5a49cdd8d nodeName:}" failed. No retries permitted until 2025-12-03 17:33:13.268393704 +0000 UTC m=+1210.085864663 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert") pod "infra-operator-controller-manager-57548d458d-phcn8" (UID: "b514a6d2-89f9-451c-af24-c0c5a49cdd8d") : secret "infra-operator-webhook-server-cert" not found Dec 03 17:32:57 crc kubenswrapper[4787]: E1203 17:32:57.324107 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5" Dec 03 17:32:57 crc kubenswrapper[4787]: E1203 17:32:57.324414 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/horizon-operator@sha256:9e847f4dbdea19ab997f32a02b3680a9bd966f9c705911645c3866a19fda9ea5,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9th54,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-operator-controller-manager-68c6d99b8f-cjbzf_openstack-operators(2a789aa2-cc4a-42a9-b1c0-9d1a648fd509): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:32:57 crc kubenswrapper[4787]: E1203 17:32:57.887508 4787 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:57 crc kubenswrapper[4787]: E1203 17:32:57.887580 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert podName:31c5513f-61f7-4d6c-8136-b212cc171732 nodeName:}" failed. No retries permitted until 2025-12-03 17:33:13.88755616 +0000 UTC m=+1210.705027119 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert") pod "openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" (UID: "31c5513f-61f7-4d6c-8136-b212cc171732") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 03 17:32:57 crc kubenswrapper[4787]: I1203 17:32:57.886568 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:32:58 crc kubenswrapper[4787]: I1203 17:32:58.296137 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:58 crc kubenswrapper[4787]: I1203 17:32:58.296193 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:32:58 crc kubenswrapper[4787]: E1203 17:32:58.296383 4787 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 03 17:32:58 crc kubenswrapper[4787]: E1203 17:32:58.296430 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:33:14.296414334 +0000 UTC m=+1211.113885293 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "webhook-server-cert" not found Dec 03 17:32:58 crc kubenswrapper[4787]: E1203 17:32:58.297069 4787 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 03 17:32:58 crc kubenswrapper[4787]: E1203 17:32:58.297098 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs podName:d9b27877-9ba0-4487-a487-aceafca2075a nodeName:}" failed. No retries permitted until 2025-12-03 17:33:14.297089632 +0000 UTC m=+1211.114560591 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs") pod "openstack-operator-controller-manager-5574b8467d-tg2v7" (UID: "d9b27877-9ba0-4487-a487-aceafca2075a") : secret "metrics-server-cert" not found Dec 03 17:33:00 crc kubenswrapper[4787]: E1203 17:33:00.362540 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670" Dec 03 17:33:00 crc kubenswrapper[4787]: E1203 17:33:00.363071 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-52vmt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-zwmxn_openstack-operators(a0326a22-7690-48ed-84f5-01cc9e8331b2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:33:01 crc kubenswrapper[4787]: E1203 17:33:01.123398 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 03 17:33:01 crc kubenswrapper[4787]: E1203 17:33:01.123839 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-gcvms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-6txhw_openstack-operators(f94890f3-3499-483d-9cda-24a377e982a8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:33:01 crc kubenswrapper[4787]: E1203 17:33:01.835145 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 03 17:33:01 crc kubenswrapper[4787]: E1203 17:33:01.835314 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-hsndp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-cnjbp_openstack-operators(c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:33:01 crc kubenswrapper[4787]: E1203 17:33:01.839064 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp" podUID="c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1" Dec 03 17:33:02 crc kubenswrapper[4787]: E1203 17:33:02.812961 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp" podUID="c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1" Dec 03 17:33:09 crc kubenswrapper[4787]: I1203 17:33:09.881332 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" event={"ID":"d1364502-c591-47c8-b24c-00c77e3a8d23","Type":"ContainerStarted","Data":"7ad3764c350dd860de933dba1f1fbfc962497d831e68ed2b24612148a61f7fc8"} Dec 03 17:33:09 crc kubenswrapper[4787]: I1203 17:33:09.892935 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg" event={"ID":"8c4a1981-21a3-49ec-a82d-419f8c080fde","Type":"ContainerStarted","Data":"225243062a652ac42296a595dc798e665c8d2eddad4dfc83bb07672b87fc461b"} Dec 03 17:33:09 crc kubenswrapper[4787]: I1203 17:33:09.895047 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk" event={"ID":"a6f58c54-821e-4f2b-bce7-1750f64f7ed9","Type":"ContainerStarted","Data":"47873af38787f8d93d1809235f9930fb29620a5a4ba5cb265c5de2cb2dd9fb2c"} Dec 03 17:33:09 crc kubenswrapper[4787]: I1203 17:33:09.898368 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7" event={"ID":"6d12a43d-d3da-4b99-b48b-519d660d2527","Type":"ContainerStarted","Data":"1b271679df82ca7a25f8bf67db6e2ccba3058667ea73dbf3e3a6ef98a3a61802"} Dec 03 17:33:09 crc kubenswrapper[4787]: I1203 17:33:09.901853 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n" event={"ID":"355352cc-c3f1-4430-8ff0-55777890e135","Type":"ContainerStarted","Data":"849b5e105792b3abf0a3dfb37d99d6fee63c3fe7407c325ad7986a3d12265cce"} Dec 03 17:33:09 crc kubenswrapper[4787]: I1203 17:33:09.903932 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-75r2j" event={"ID":"67472bf9-12b8-4463-bc95-dec19e689f36","Type":"ContainerStarted","Data":"14f608ea70fc7c8a977b72010e4e317d98f66686e2e91da4cdbbf6346b6578eb"} Dec 03 17:33:09 crc kubenswrapper[4787]: I1203 17:33:09.905806 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr" event={"ID":"226cdcb0-0abc-4a06-9d5c-1d30d9ac8288","Type":"ContainerStarted","Data":"6251ee3131318e92c67017587577bfb9a32d1d4186c8d1f726817c5a48385ad4"} Dec 03 17:33:09 crc kubenswrapper[4787]: I1203 17:33:09.923665 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn" event={"ID":"43962cfc-342c-49db-83bf-ccde92708a0b","Type":"ContainerStarted","Data":"d466fd34a4bcfe610365cf45219797fdbe53f50c2de96eb85aa620e5e2162a21"} Dec 03 17:33:10 crc kubenswrapper[4787]: E1203 17:33:10.199625 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5ljcj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod neutron-operator-controller-manager-5fdfd5b6b5-sbcbk_openstack-operators(04fbeb64-47f2-4393-8295-0ac7d6db4df3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:33:10 crc kubenswrapper[4787]: E1203 17:33:10.200387 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-klqkr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f8c65bbfc-vmpfz_openstack-operators(c209ad3f-9014-4385-96cb-323dd27b8335): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 03 17:33:10 crc kubenswrapper[4787]: E1203 17:33:10.201639 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" podUID="c209ad3f-9014-4385-96cb-323dd27b8335" Dec 03 17:33:10 crc kubenswrapper[4787]: E1203 17:33:10.201706 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" podUID="04fbeb64-47f2-4393-8295-0ac7d6db4df3" Dec 03 17:33:10 crc kubenswrapper[4787]: I1203 17:33:10.954872 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" event={"ID":"5c7f6976-b7e2-453d-ad11-6d0e978cd03d","Type":"ContainerStarted","Data":"8a33486b5301dcfacfcd0c556b738b457185e870c6fc674f3a0f976476a9f231"} Dec 03 17:33:10 crc kubenswrapper[4787]: I1203 17:33:10.963849 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" event={"ID":"8af08022-68dd-4fc4-a052-2cfde5f1ab0a","Type":"ContainerStarted","Data":"e2549404fd99533b361d6df448d8210afd0d5641e2141d20ef99e2a688de4c1d"} Dec 03 17:33:10 crc kubenswrapper[4787]: I1203 17:33:10.979318 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8" event={"ID":"dd3a3971-adb9-46cb-a0b7-63897e326341","Type":"ContainerStarted","Data":"9d374bf260fe54fd983ba545838babafbf9c0ffee01d053d305d8a24d440f40c"} Dec 03 17:33:10 crc kubenswrapper[4787]: I1203 17:33:10.983280 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" event={"ID":"c209ad3f-9014-4385-96cb-323dd27b8335","Type":"ContainerStarted","Data":"c9940d56a86f13b9f802c38d56220da02a1bd30c1ec4a04bd28a5b03f281d008"} Dec 03 17:33:10 crc kubenswrapper[4787]: I1203 17:33:10.983561 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" Dec 03 17:33:10 crc kubenswrapper[4787]: I1203 17:33:10.984631 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2" event={"ID":"a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173","Type":"ContainerStarted","Data":"c0f25e5fd1fef8917b4605298aa3c4eabd6fb68ce1b2fce700058cfcc0695a0e"} Dec 03 17:33:10 crc kubenswrapper[4787]: E1203 17:33:10.985350 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" podUID="c209ad3f-9014-4385-96cb-323dd27b8335" Dec 03 17:33:10 crc kubenswrapper[4787]: I1203 17:33:10.992719 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" event={"ID":"04fbeb64-47f2-4393-8295-0ac7d6db4df3","Type":"ContainerStarted","Data":"cdb817a72b1927511559f95c965169e7da4efdb0be749755ee1484d51c45d269"} Dec 03 17:33:10 crc kubenswrapper[4787]: I1203 17:33:10.992939 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" Dec 03 17:33:10 crc kubenswrapper[4787]: E1203 17:33:10.994610 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" podUID="04fbeb64-47f2-4393-8295-0ac7d6db4df3" Dec 03 17:33:12 crc kubenswrapper[4787]: E1203 17:33:12.002090 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" podUID="04fbeb64-47f2-4393-8295-0ac7d6db4df3" Dec 03 17:33:12 crc kubenswrapper[4787]: E1203 17:33:12.002090 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" podUID="c209ad3f-9014-4385-96cb-323dd27b8335" Dec 03 17:33:13 crc kubenswrapper[4787]: I1203 17:33:13.273349 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:33:13 crc kubenswrapper[4787]: I1203 17:33:13.281755 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/b514a6d2-89f9-451c-af24-c0c5a49cdd8d-cert\") pod \"infra-operator-controller-manager-57548d458d-phcn8\" (UID: \"b514a6d2-89f9-451c-af24-c0c5a49cdd8d\") " pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:33:13 crc kubenswrapper[4787]: I1203 17:33:13.548416 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-rfvjr" Dec 03 17:33:13 crc kubenswrapper[4787]: I1203 17:33:13.556620 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:33:13 crc kubenswrapper[4787]: I1203 17:33:13.986182 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:33:13 crc kubenswrapper[4787]: I1203 17:33:13.990511 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/31c5513f-61f7-4d6c-8136-b212cc171732-cert\") pod \"openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz\" (UID: \"31c5513f-61f7-4d6c-8136-b212cc171732\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:33:14 crc kubenswrapper[4787]: I1203 17:33:14.256922 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-7l5tk" Dec 03 17:33:14 crc kubenswrapper[4787]: I1203 17:33:14.264730 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:33:14 crc kubenswrapper[4787]: I1203 17:33:14.397771 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:33:14 crc kubenswrapper[4787]: I1203 17:33:14.397815 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:33:14 crc kubenswrapper[4787]: I1203 17:33:14.401835 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-webhook-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:33:14 crc kubenswrapper[4787]: I1203 17:33:14.401948 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d9b27877-9ba0-4487-a487-aceafca2075a-metrics-certs\") pod \"openstack-operator-controller-manager-5574b8467d-tg2v7\" (UID: \"d9b27877-9ba0-4487-a487-aceafca2075a\") " pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:33:14 crc kubenswrapper[4787]: I1203 17:33:14.694714 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-tg82f" Dec 03 17:33:14 crc kubenswrapper[4787]: I1203 17:33:14.702297 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:33:18 crc kubenswrapper[4787]: I1203 17:33:18.989653 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:33:18 crc kubenswrapper[4787]: I1203 17:33:18.990288 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:33:22 crc kubenswrapper[4787]: I1203 17:33:22.098405 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" Dec 03 17:33:22 crc kubenswrapper[4787]: I1203 17:33:22.449776 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" Dec 03 17:33:22 crc kubenswrapper[4787]: E1203 17:33:22.451795 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" podUID="c209ad3f-9014-4385-96cb-323dd27b8335" Dec 03 17:33:23 crc kubenswrapper[4787]: I1203 17:33:23.717906 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz"] Dec 03 17:33:23 crc kubenswrapper[4787]: W1203 17:33:23.766037 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod31c5513f_61f7_4d6c_8136_b212cc171732.slice/crio-4424a19068240037d46eb64ec39769e33361c0fff5972c31e40976a89738203c WatchSource:0}: Error finding container 4424a19068240037d46eb64ec39769e33361c0fff5972c31e40976a89738203c: Status 404 returned error can't find the container with id 4424a19068240037d46eb64ec39769e33361c0fff5972c31e40976a89738203c Dec 03 17:33:23 crc kubenswrapper[4787]: E1203 17:33:23.905721 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = reading blob sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce\": context canceled" image="quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0" Dec 03 17:33:23 crc kubenswrapper[4787]: E1203 17:33:23.905918 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-wldnb,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-56bbcc9d85-5m5h7_openstack-operators(6384c305-657a-465d-8e48-487fb12cd52f): ErrImagePull: rpc error: code = Canceled desc = reading blob sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce: Get \"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce\": context canceled" logger="UnhandledError" Dec 03 17:33:23 crc kubenswrapper[4787]: E1203 17:33:23.907150 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"rpc error: code = Canceled desc = reading blob sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce: Get \\\"https://quay.io/v2/openstack-k8s-operators/kube-rbac-proxy/blobs/sha256:4fa131a1b726b2d6468d461e7d8867a2157d5671f712461d8abd126155fdf9ce\\\": context canceled\"]" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" podUID="6384c305-657a-465d-8e48-487fb12cd52f" Dec 03 17:33:23 crc kubenswrapper[4787]: I1203 17:33:23.962170 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-57548d458d-phcn8"] Dec 03 17:33:24 crc kubenswrapper[4787]: I1203 17:33:24.021781 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7"] Dec 03 17:33:24 crc kubenswrapper[4787]: I1203 17:33:24.111812 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" event={"ID":"31c5513f-61f7-4d6c-8136-b212cc171732","Type":"ContainerStarted","Data":"4424a19068240037d46eb64ec39769e33361c0fff5972c31e40976a89738203c"} Dec 03 17:33:25 crc kubenswrapper[4787]: W1203 17:33:25.287777 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb514a6d2_89f9_451c_af24_c0c5a49cdd8d.slice/crio-b94e37ee0a15f30b5edc2d57322d2a8a20d796dea9ee1d20cdbec1da8b34ff58 WatchSource:0}: Error finding container b94e37ee0a15f30b5edc2d57322d2a8a20d796dea9ee1d20cdbec1da8b34ff58: Status 404 returned error can't find the container with id b94e37ee0a15f30b5edc2d57322d2a8a20d796dea9ee1d20cdbec1da8b34ff58 Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.141225 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" event={"ID":"d1364502-c591-47c8-b24c-00c77e3a8d23","Type":"ContainerStarted","Data":"e610ffe4881d48ede14c6a75745c97fccbdff12ede7c3538615719cecad6d2af"} Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.142767 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.148237 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" event={"ID":"d9b27877-9ba0-4487-a487-aceafca2075a","Type":"ContainerStarted","Data":"4f5eda4a46015a1791e562a6ccc24bc8def0c003b1407e45001e195ef8661be8"} Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.148282 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" event={"ID":"d9b27877-9ba0-4487-a487-aceafca2075a","Type":"ContainerStarted","Data":"2583aaae2337db1f3a5bc808d67e98b622e7d377edecc61db228df1c8ca3ea68"} Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.148929 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.149139 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.160098 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg" event={"ID":"8c4a1981-21a3-49ec-a82d-419f8c080fde","Type":"ContainerStarted","Data":"7d804190e8e4b66ccc0787bce6bec19846b2595f4187aed83947bde953ed6be6"} Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.160694 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.163764 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp" event={"ID":"c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1","Type":"ContainerStarted","Data":"076685edec6345325cc936383909da0aa5c5a9111f899e9f4ba6eb9f65317b58"} Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.166758 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-wtvd4" podStartSLOduration=3.405532333 podStartE2EDuration="45.166739572s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.799047324 +0000 UTC m=+1180.616518283" lastFinishedPulling="2025-12-03 17:33:25.560254563 +0000 UTC m=+1222.377725522" observedRunningTime="2025-12-03 17:33:26.163505487 +0000 UTC m=+1222.980976446" watchObservedRunningTime="2025-12-03 17:33:26.166739572 +0000 UTC m=+1222.984210531" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.169243 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.173319 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" event={"ID":"b514a6d2-89f9-451c-af24-c0c5a49cdd8d","Type":"ContainerStarted","Data":"b94e37ee0a15f30b5edc2d57322d2a8a20d796dea9ee1d20cdbec1da8b34ff58"} Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.177260 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" event={"ID":"6384c305-657a-465d-8e48-487fb12cd52f","Type":"ContainerStarted","Data":"c42919b0c03e4d230889bb1b5b6badeaf3166fad58a49ee8273e2e0fcd1fba17"} Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.189071 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.193230 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" event={"ID":"8af08022-68dd-4fc4-a052-2cfde5f1ab0a","Type":"ContainerStarted","Data":"ecc4c69726f46a1cb521cda645a21fbb5103ed28f69d3fbd3643b14a4cf97f43"} Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.194355 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.195040 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.201642 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.227361 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-6c548fd776-bgmsg" podStartSLOduration=3.028904813 podStartE2EDuration="45.227345315s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.461202599 +0000 UTC m=+1180.278673558" lastFinishedPulling="2025-12-03 17:33:25.659643111 +0000 UTC m=+1222.477114060" observedRunningTime="2025-12-03 17:33:26.224822199 +0000 UTC m=+1223.042293158" watchObservedRunningTime="2025-12-03 17:33:26.227345315 +0000 UTC m=+1223.044816274" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.309588 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" podStartSLOduration=45.30957129 podStartE2EDuration="45.30957129s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:33:26.273848815 +0000 UTC m=+1223.091319774" watchObservedRunningTime="2025-12-03 17:33:26.30957129 +0000 UTC m=+1223.127042249" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.313212 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-cnjbp" podStartSLOduration=5.605410047 podStartE2EDuration="45.313203746s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.766520774 +0000 UTC m=+1180.583991733" lastFinishedPulling="2025-12-03 17:33:23.474314473 +0000 UTC m=+1220.291785432" observedRunningTime="2025-12-03 17:33:26.311610444 +0000 UTC m=+1223.129081413" watchObservedRunningTime="2025-12-03 17:33:26.313203746 +0000 UTC m=+1223.130674715" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.338400 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-769dc69bc-2wq6s" podStartSLOduration=3.646639351 podStartE2EDuration="45.338376912s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.794809002 +0000 UTC m=+1180.612279951" lastFinishedPulling="2025-12-03 17:33:25.486546553 +0000 UTC m=+1222.304017512" observedRunningTime="2025-12-03 17:33:26.332539898 +0000 UTC m=+1223.150010857" watchObservedRunningTime="2025-12-03 17:33:26.338376912 +0000 UTC m=+1223.155847871" Dec 03 17:33:26 crc kubenswrapper[4787]: I1203 17:33:26.400636 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr" podStartSLOduration=3.209748876 podStartE2EDuration="45.400618708s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.464309851 +0000 UTC m=+1180.281780810" lastFinishedPulling="2025-12-03 17:33:25.655179683 +0000 UTC m=+1222.472650642" observedRunningTime="2025-12-03 17:33:26.394456555 +0000 UTC m=+1223.211927514" watchObservedRunningTime="2025-12-03 17:33:26.400618708 +0000 UTC m=+1223.218089667" Dec 03 17:33:26 crc kubenswrapper[4787]: E1203 17:33:26.762507 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" podUID="a0326a22-7690-48ed-84f5-01cc9e8331b2" Dec 03 17:33:26 crc kubenswrapper[4787]: E1203 17:33:26.765716 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" podUID="2a789aa2-cc4a-42a9-b1c0-9d1a648fd509" Dec 03 17:33:26 crc kubenswrapper[4787]: E1203 17:33:26.768399 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" podUID="f94890f3-3499-483d-9cda-24a377e982a8" Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.206564 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-xg7kr" event={"ID":"226cdcb0-0abc-4a06-9d5c-1d30d9ac8288","Type":"ContainerStarted","Data":"fd11963dff3bf5ad81b23fe89999135433833c6d60a133784b148c6f50e2ab5f"} Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.211509 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" event={"ID":"5c7f6976-b7e2-453d-ad11-6d0e978cd03d","Type":"ContainerStarted","Data":"e8264773f117d19fdc854193c881c34f23f0cf1788980cb688c1d142115fbfc7"} Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.212257 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.217256 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.218338 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" event={"ID":"2a789aa2-cc4a-42a9-b1c0-9d1a648fd509","Type":"ContainerStarted","Data":"b2dfc6987c328b99a65c45f2c452647252b77a694e1c66355db1d0b44fb16ee3"} Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.225808 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" event={"ID":"04fbeb64-47f2-4393-8295-0ac7d6db4df3","Type":"ContainerStarted","Data":"d462eb3f8ddccd8a88cf14e47d9f116f1d8334ff09380f4888c11297ef25934c"} Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.227398 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" event={"ID":"f94890f3-3499-483d-9cda-24a377e982a8","Type":"ContainerStarted","Data":"8256382798d8bb776b0f2d1638a3aac2c031b79151f367a7d3ce20f224200cdd"} Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.228473 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" event={"ID":"a0326a22-7690-48ed-84f5-01cc9e8331b2","Type":"ContainerStarted","Data":"7205d94c5e13bf1ea315fa94009f4ec0e2ce4c5493315cea264c935f24e493b8"} Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.235487 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7" event={"ID":"6d12a43d-d3da-4b99-b48b-519d660d2527","Type":"ContainerStarted","Data":"672fd5b96000a3e6ec9c43e3cde1e2f47fb0edebfa1a46e9c332b9d285fc4125"} Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.235574 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7" Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.239667 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7" Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.280959 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-859b6ccc6-fgdm7" podStartSLOduration=4.606499379 podStartE2EDuration="47.280939191s" podCreationTimestamp="2025-12-03 17:32:40 +0000 UTC" firstStartedPulling="2025-12-03 17:32:42.989456142 +0000 UTC m=+1179.806927101" lastFinishedPulling="2025-12-03 17:33:25.663895954 +0000 UTC m=+1222.481366913" observedRunningTime="2025-12-03 17:33:27.271529252 +0000 UTC m=+1224.089000211" watchObservedRunningTime="2025-12-03 17:33:27.280939191 +0000 UTC m=+1224.098410150" Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.281240 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-7c79b5df47-cd9jr" podStartSLOduration=4.544574599 podStartE2EDuration="46.281236329s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.804725434 +0000 UTC m=+1180.622196393" lastFinishedPulling="2025-12-03 17:33:25.541387164 +0000 UTC m=+1222.358858123" observedRunningTime="2025-12-03 17:33:27.248668818 +0000 UTC m=+1224.066139777" watchObservedRunningTime="2025-12-03 17:33:27.281236329 +0000 UTC m=+1224.098707288" Dec 03 17:33:27 crc kubenswrapper[4787]: E1203 17:33:27.396160 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" podUID="02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23" Dec 03 17:33:27 crc kubenswrapper[4787]: I1203 17:33:27.492609 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-sbcbk" podStartSLOduration=4.285246741 podStartE2EDuration="46.492589169s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.464300641 +0000 UTC m=+1180.281771600" lastFinishedPulling="2025-12-03 17:33:25.671643069 +0000 UTC m=+1222.489114028" observedRunningTime="2025-12-03 17:33:27.481926287 +0000 UTC m=+1224.299397276" watchObservedRunningTime="2025-12-03 17:33:27.492589169 +0000 UTC m=+1224.310060128" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.244746 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n" event={"ID":"355352cc-c3f1-4430-8ff0-55777890e135","Type":"ContainerStarted","Data":"d5a630d39e2637ea84c0994adf3c9b4a07c7fc6a715ff4536ac5f6311ab72d00"} Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.245403 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.247706 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-75r2j" event={"ID":"67472bf9-12b8-4463-bc95-dec19e689f36","Type":"ContainerStarted","Data":"c0d6a5e08306ea6e412d0a661e471219af16094ab6b0da03ec453c9938653c2d"} Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.247918 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-75r2j" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.249464 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.249688 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-75r2j" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.250389 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" event={"ID":"02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23","Type":"ContainerStarted","Data":"1d2e1a160894ed89c134774bd9b0440af41a11dbd54ee881846db582183d3ac8"} Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.255327 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn" event={"ID":"43962cfc-342c-49db-83bf-ccde92708a0b","Type":"ContainerStarted","Data":"f303a619538299bb411bbbfac92f491c9a65aedbbf4e7fb33c1bdbd92d26e65e"} Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.256142 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.257228 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.258801 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8" event={"ID":"dd3a3971-adb9-46cb-a0b7-63897e326341","Type":"ContainerStarted","Data":"36b3be4d7c80ab31d15b419330e32f84896d279f43d038309308a61816309d23"} Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.259095 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.260513 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.268711 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2" event={"ID":"a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173","Type":"ContainerStarted","Data":"f88f7c49073071df76f64ecc07e0db495f7e27764ae3d3528e76e0059a46060f"} Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.269478 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.271700 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.273520 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk" event={"ID":"a6f58c54-821e-4f2b-bce7-1750f64f7ed9","Type":"ContainerStarted","Data":"b8f518de6be88135cf671cc3adb6609e6a131935d4a2b59b3b04706b2b8739a3"} Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.274368 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.274449 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-77987cd8cd-p425n" podStartSLOduration=5.441378402 podStartE2EDuration="48.274428658s" podCreationTimestamp="2025-12-03 17:32:40 +0000 UTC" firstStartedPulling="2025-12-03 17:32:42.826579585 +0000 UTC m=+1179.644050544" lastFinishedPulling="2025-12-03 17:33:25.659629841 +0000 UTC m=+1222.477100800" observedRunningTime="2025-12-03 17:33:28.265476141 +0000 UTC m=+1225.082947120" watchObservedRunningTime="2025-12-03 17:33:28.274428658 +0000 UTC m=+1225.091899617" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.276904 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.283467 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" event={"ID":"6384c305-657a-465d-8e48-487fb12cd52f","Type":"ContainerStarted","Data":"75e7a046dbfbdb13df2e64c284bfa2876981b34d90922a5464fb1ac5f1b64f44"} Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.283704 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.294333 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-75r2j" podStartSLOduration=5.423979809 podStartE2EDuration="47.294312084s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.79547675 +0000 UTC m=+1180.612947709" lastFinishedPulling="2025-12-03 17:33:25.665809025 +0000 UTC m=+1222.483279984" observedRunningTime="2025-12-03 17:33:28.279712597 +0000 UTC m=+1225.097183556" watchObservedRunningTime="2025-12-03 17:33:28.294312084 +0000 UTC m=+1225.111783043" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.321270 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-bzjxn" podStartSLOduration=5.207195828 podStartE2EDuration="48.321246266s" podCreationTimestamp="2025-12-03 17:32:40 +0000 UTC" firstStartedPulling="2025-12-03 17:32:42.549771014 +0000 UTC m=+1179.367241973" lastFinishedPulling="2025-12-03 17:33:25.663821452 +0000 UTC m=+1222.481292411" observedRunningTime="2025-12-03 17:33:28.302693705 +0000 UTC m=+1225.120164664" watchObservedRunningTime="2025-12-03 17:33:28.321246266 +0000 UTC m=+1225.138717225" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.372855 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-zgkn8" podStartSLOduration=5.476800895 podStartE2EDuration="47.3728325s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.763665218 +0000 UTC m=+1180.581136177" lastFinishedPulling="2025-12-03 17:33:25.659696823 +0000 UTC m=+1222.477167782" observedRunningTime="2025-12-03 17:33:28.367252043 +0000 UTC m=+1225.184723012" watchObservedRunningTime="2025-12-03 17:33:28.3728325 +0000 UTC m=+1225.190303459" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.408562 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-66f6f55998-k4kg2" podStartSLOduration=5.511071502 podStartE2EDuration="47.408543295s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.766444552 +0000 UTC m=+1180.583915511" lastFinishedPulling="2025-12-03 17:33:25.663916345 +0000 UTC m=+1222.481387304" observedRunningTime="2025-12-03 17:33:28.386643096 +0000 UTC m=+1225.204114075" watchObservedRunningTime="2025-12-03 17:33:28.408543295 +0000 UTC m=+1225.226014254" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.430364 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" podStartSLOduration=5.257206618 podStartE2EDuration="47.430335181s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.493200955 +0000 UTC m=+1180.310671914" lastFinishedPulling="2025-12-03 17:33:25.666329518 +0000 UTC m=+1222.483800477" observedRunningTime="2025-12-03 17:33:28.405551116 +0000 UTC m=+1225.223022105" watchObservedRunningTime="2025-12-03 17:33:28.430335181 +0000 UTC m=+1225.247806140" Dec 03 17:33:28 crc kubenswrapper[4787]: I1203 17:33:28.447228 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-hr6rk" podStartSLOduration=5.567792254 podStartE2EDuration="48.447205967s" podCreationTimestamp="2025-12-03 17:32:40 +0000 UTC" firstStartedPulling="2025-12-03 17:32:42.996783525 +0000 UTC m=+1179.814254474" lastFinishedPulling="2025-12-03 17:33:25.876197228 +0000 UTC m=+1222.693668187" observedRunningTime="2025-12-03 17:33:28.424463636 +0000 UTC m=+1225.241934595" watchObservedRunningTime="2025-12-03 17:33:28.447205967 +0000 UTC m=+1225.264676926" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.301811 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" event={"ID":"f94890f3-3499-483d-9cda-24a377e982a8","Type":"ContainerStarted","Data":"2ed91e49f96920ce56d8d089f824de3b2495d982359a556460d8d64ff531e438"} Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.302400 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.304301 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" event={"ID":"b514a6d2-89f9-451c-af24-c0c5a49cdd8d","Type":"ContainerStarted","Data":"7108f49ae07d55e0bb952772a09779dc1eb308693aac27065a74f4c9ec7b04db"} Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.304343 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" event={"ID":"b514a6d2-89f9-451c-af24-c0c5a49cdd8d","Type":"ContainerStarted","Data":"6ae3738aae2dd98cf260237584b1ac1ba1a9581a7b68a17712a8c692b0dd22b3"} Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.304455 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.310198 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" event={"ID":"a0326a22-7690-48ed-84f5-01cc9e8331b2","Type":"ContainerStarted","Data":"a65704e4cd2a003bab2d9c12fa7ff3e3c6b0557bbca9ce764d660fe5f8361e42"} Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.310963 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.316652 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" event={"ID":"02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23","Type":"ContainerStarted","Data":"ed3550f11b6686bf970ab663a77204f5ba099d81118c5c3f6ea06f996efe2580"} Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.320141 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.341490 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" event={"ID":"2a789aa2-cc4a-42a9-b1c0-9d1a648fd509","Type":"ContainerStarted","Data":"c44503b8675643ab655c95ac18292da35faf87149c9d78791b2e5307070a3ed0"} Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.342378 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.343650 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" podStartSLOduration=3.649301781 podStartE2EDuration="49.343629584s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.565675732 +0000 UTC m=+1180.383146691" lastFinishedPulling="2025-12-03 17:33:29.260003535 +0000 UTC m=+1226.077474494" observedRunningTime="2025-12-03 17:33:30.338757825 +0000 UTC m=+1227.156228784" watchObservedRunningTime="2025-12-03 17:33:30.343629584 +0000 UTC m=+1227.161100543" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.351522 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" event={"ID":"31c5513f-61f7-4d6c-8136-b212cc171732","Type":"ContainerStarted","Data":"1882aed7f33141227ba4dc96be94aa5b3bdcab0ee484f56e25fdedf597d3c963"} Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.351569 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" event={"ID":"31c5513f-61f7-4d6c-8136-b212cc171732","Type":"ContainerStarted","Data":"784d5e7c1b712f40897d1f8435cb486fd35911e482cce2d81b70c1a439069a7c"} Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.351586 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.371187 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" podStartSLOduration=3.900263301 podStartE2EDuration="50.371171762s" podCreationTimestamp="2025-12-03 17:32:40 +0000 UTC" firstStartedPulling="2025-12-03 17:32:42.795324078 +0000 UTC m=+1179.612795037" lastFinishedPulling="2025-12-03 17:33:29.266232529 +0000 UTC m=+1226.083703498" observedRunningTime="2025-12-03 17:33:30.369505218 +0000 UTC m=+1227.186976177" watchObservedRunningTime="2025-12-03 17:33:30.371171762 +0000 UTC m=+1227.188642711" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.390196 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" podStartSLOduration=45.437306899 podStartE2EDuration="49.390181765s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:33:25.31437302 +0000 UTC m=+1222.131843979" lastFinishedPulling="2025-12-03 17:33:29.267247886 +0000 UTC m=+1226.084718845" observedRunningTime="2025-12-03 17:33:30.387817523 +0000 UTC m=+1227.205288472" watchObservedRunningTime="2025-12-03 17:33:30.390181765 +0000 UTC m=+1227.207652714" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.407379 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" podStartSLOduration=3.148068084 podStartE2EDuration="49.407329619s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.000812112 +0000 UTC m=+1179.818283071" lastFinishedPulling="2025-12-03 17:33:29.260073607 +0000 UTC m=+1226.077544606" observedRunningTime="2025-12-03 17:33:30.400617411 +0000 UTC m=+1227.218088370" watchObservedRunningTime="2025-12-03 17:33:30.407329619 +0000 UTC m=+1227.224800578" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.429276 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" podStartSLOduration=3.735246274 podStartE2EDuration="49.429256709s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.566479113 +0000 UTC m=+1180.383950072" lastFinishedPulling="2025-12-03 17:33:29.260489548 +0000 UTC m=+1226.077960507" observedRunningTime="2025-12-03 17:33:30.427579984 +0000 UTC m=+1227.245050943" watchObservedRunningTime="2025-12-03 17:33:30.429256709 +0000 UTC m=+1227.246727678" Dec 03 17:33:30 crc kubenswrapper[4787]: I1203 17:33:30.458269 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" podStartSLOduration=43.968493451 podStartE2EDuration="49.458254066s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:33:23.76986716 +0000 UTC m=+1220.587338119" lastFinishedPulling="2025-12-03 17:33:29.259627765 +0000 UTC m=+1226.077098734" observedRunningTime="2025-12-03 17:33:30.455856812 +0000 UTC m=+1227.273327771" watchObservedRunningTime="2025-12-03 17:33:30.458254066 +0000 UTC m=+1227.275725025" Dec 03 17:33:32 crc kubenswrapper[4787]: I1203 17:33:32.047166 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-56bbcc9d85-5m5h7" Dec 03 17:33:34 crc kubenswrapper[4787]: I1203 17:33:34.272655 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz" Dec 03 17:33:34 crc kubenswrapper[4787]: I1203 17:33:34.711099 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-5574b8467d-tg2v7" Dec 03 17:33:37 crc kubenswrapper[4787]: I1203 17:33:37.409309 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" event={"ID":"c209ad3f-9014-4385-96cb-323dd27b8335","Type":"ContainerStarted","Data":"c370306566f9e377f28321252cdded074ad18b656574be5fac06204a7ec384c5"} Dec 03 17:33:37 crc kubenswrapper[4787]: I1203 17:33:37.435544 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f8c65bbfc-vmpfz" podStartSLOduration=31.002167914 podStartE2EDuration="56.435519702s" podCreationTimestamp="2025-12-03 17:32:41 +0000 UTC" firstStartedPulling="2025-12-03 17:32:43.804615811 +0000 UTC m=+1180.622086760" lastFinishedPulling="2025-12-03 17:33:09.237967589 +0000 UTC m=+1206.055438548" observedRunningTime="2025-12-03 17:33:37.4286166 +0000 UTC m=+1234.246087599" watchObservedRunningTime="2025-12-03 17:33:37.435519702 +0000 UTC m=+1234.252990671" Dec 03 17:33:41 crc kubenswrapper[4787]: I1203 17:33:41.358947 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-78b4bc895b-mwpr8" Dec 03 17:33:41 crc kubenswrapper[4787]: I1203 17:33:41.836332 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-cjbzf" Dec 03 17:33:41 crc kubenswrapper[4787]: I1203 17:33:41.880758 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-zwmxn" Dec 03 17:33:41 crc kubenswrapper[4787]: I1203 17:33:41.947288 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-6txhw" Dec 03 17:33:43 crc kubenswrapper[4787]: I1203 17:33:43.566487 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-57548d458d-phcn8" Dec 03 17:33:48 crc kubenswrapper[4787]: I1203 17:33:48.990075 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:33:48 crc kubenswrapper[4787]: I1203 17:33:48.990578 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.561786 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd2cd"] Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.564406 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.568221 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.568383 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.568455 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-lsnpm" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.568580 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.572351 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd2cd"] Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.680185 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qmdcv"] Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.682680 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.693676 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qmdcv"] Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.694686 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.711552 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85be111d-8722-48ec-8f70-6c8c53c0fcd8-config\") pod \"dnsmasq-dns-675f4bcbfc-kd2cd\" (UID: \"85be111d-8722-48ec-8f70-6c8c53c0fcd8\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.711635 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-495zj\" (UniqueName: \"kubernetes.io/projected/85be111d-8722-48ec-8f70-6c8c53c0fcd8-kube-api-access-495zj\") pod \"dnsmasq-dns-675f4bcbfc-kd2cd\" (UID: \"85be111d-8722-48ec-8f70-6c8c53c0fcd8\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.814392 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-config\") pod \"dnsmasq-dns-78dd6ddcc-qmdcv\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.814466 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85be111d-8722-48ec-8f70-6c8c53c0fcd8-config\") pod \"dnsmasq-dns-675f4bcbfc-kd2cd\" (UID: \"85be111d-8722-48ec-8f70-6c8c53c0fcd8\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.814526 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-495zj\" (UniqueName: \"kubernetes.io/projected/85be111d-8722-48ec-8f70-6c8c53c0fcd8-kube-api-access-495zj\") pod \"dnsmasq-dns-675f4bcbfc-kd2cd\" (UID: \"85be111d-8722-48ec-8f70-6c8c53c0fcd8\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.814571 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nr6rl\" (UniqueName: \"kubernetes.io/projected/e231ba27-59a3-48a6-b716-cb1a18a07602-kube-api-access-nr6rl\") pod \"dnsmasq-dns-78dd6ddcc-qmdcv\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.814602 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-qmdcv\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.815664 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85be111d-8722-48ec-8f70-6c8c53c0fcd8-config\") pod \"dnsmasq-dns-675f4bcbfc-kd2cd\" (UID: \"85be111d-8722-48ec-8f70-6c8c53c0fcd8\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.853346 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-495zj\" (UniqueName: \"kubernetes.io/projected/85be111d-8722-48ec-8f70-6c8c53c0fcd8-kube-api-access-495zj\") pod \"dnsmasq-dns-675f4bcbfc-kd2cd\" (UID: \"85be111d-8722-48ec-8f70-6c8c53c0fcd8\") " pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.887304 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.916565 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-qmdcv\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.917023 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nr6rl\" (UniqueName: \"kubernetes.io/projected/e231ba27-59a3-48a6-b716-cb1a18a07602-kube-api-access-nr6rl\") pod \"dnsmasq-dns-78dd6ddcc-qmdcv\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.917352 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-config\") pod \"dnsmasq-dns-78dd6ddcc-qmdcv\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.918162 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-qmdcv\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.918646 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-config\") pod \"dnsmasq-dns-78dd6ddcc-qmdcv\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:00 crc kubenswrapper[4787]: I1203 17:34:00.936059 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nr6rl\" (UniqueName: \"kubernetes.io/projected/e231ba27-59a3-48a6-b716-cb1a18a07602-kube-api-access-nr6rl\") pod \"dnsmasq-dns-78dd6ddcc-qmdcv\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:01 crc kubenswrapper[4787]: I1203 17:34:01.002730 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:01 crc kubenswrapper[4787]: I1203 17:34:01.382548 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd2cd"] Dec 03 17:34:01 crc kubenswrapper[4787]: I1203 17:34:01.501725 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qmdcv"] Dec 03 17:34:01 crc kubenswrapper[4787]: W1203 17:34:01.503766 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode231ba27_59a3_48a6_b716_cb1a18a07602.slice/crio-16713232cca87da1c8982a1d9811f2a676dd1ff00933d9807f0da473b86805a8 WatchSource:0}: Error finding container 16713232cca87da1c8982a1d9811f2a676dd1ff00933d9807f0da473b86805a8: Status 404 returned error can't find the container with id 16713232cca87da1c8982a1d9811f2a676dd1ff00933d9807f0da473b86805a8 Dec 03 17:34:01 crc kubenswrapper[4787]: I1203 17:34:01.619694 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" event={"ID":"85be111d-8722-48ec-8f70-6c8c53c0fcd8","Type":"ContainerStarted","Data":"93790da9084107b7a563c441b2ab0263c59c9bf1d65d75a601d1b65f3ced224e"} Dec 03 17:34:01 crc kubenswrapper[4787]: I1203 17:34:01.623917 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" event={"ID":"e231ba27-59a3-48a6-b716-cb1a18a07602","Type":"ContainerStarted","Data":"16713232cca87da1c8982a1d9811f2a676dd1ff00933d9807f0da473b86805a8"} Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.598730 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd2cd"] Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.644563 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-dxzmh"] Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.646497 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.670142 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-dxzmh"] Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.766699 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nw9ms\" (UniqueName: \"kubernetes.io/projected/5dc53172-a8d8-44f8-9969-d5ed647dcd71-kube-api-access-nw9ms\") pod \"dnsmasq-dns-5ccc8479f9-dxzmh\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.766854 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-dxzmh\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.766987 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-config\") pod \"dnsmasq-dns-5ccc8479f9-dxzmh\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.843518 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qmdcv"] Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.868409 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nw9ms\" (UniqueName: \"kubernetes.io/projected/5dc53172-a8d8-44f8-9969-d5ed647dcd71-kube-api-access-nw9ms\") pod \"dnsmasq-dns-5ccc8479f9-dxzmh\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.868462 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-dxzmh\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.868506 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-config\") pod \"dnsmasq-dns-5ccc8479f9-dxzmh\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.871668 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-dxzmh\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.873963 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkgx"] Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.878595 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-config\") pod \"dnsmasq-dns-5ccc8479f9-dxzmh\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.899505 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.901263 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nw9ms\" (UniqueName: \"kubernetes.io/projected/5dc53172-a8d8-44f8-9969-d5ed647dcd71-kube-api-access-nw9ms\") pod \"dnsmasq-dns-5ccc8479f9-dxzmh\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.913661 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkgx"] Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.972386 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bmkgx\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.972840 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bl59w\" (UniqueName: \"kubernetes.io/projected/bd624460-116c-4630-b418-788f93ede485-kube-api-access-bl59w\") pod \"dnsmasq-dns-57d769cc4f-bmkgx\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.972930 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-config\") pod \"dnsmasq-dns-57d769cc4f-bmkgx\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:02 crc kubenswrapper[4787]: I1203 17:34:02.979845 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.075380 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bmkgx\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.075849 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bl59w\" (UniqueName: \"kubernetes.io/projected/bd624460-116c-4630-b418-788f93ede485-kube-api-access-bl59w\") pod \"dnsmasq-dns-57d769cc4f-bmkgx\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.075881 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-config\") pod \"dnsmasq-dns-57d769cc4f-bmkgx\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.078524 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-bmkgx\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.085278 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-config\") pod \"dnsmasq-dns-57d769cc4f-bmkgx\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.098171 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bl59w\" (UniqueName: \"kubernetes.io/projected/bd624460-116c-4630-b418-788f93ede485-kube-api-access-bl59w\") pod \"dnsmasq-dns-57d769cc4f-bmkgx\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.277132 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.663920 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-dxzmh"] Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.689309 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.692916 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.698353 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.698560 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.698779 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.698963 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.699194 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.699352 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.699364 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6rzx9" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.703687 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.722631 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" event={"ID":"5dc53172-a8d8-44f8-9969-d5ed647dcd71","Type":"ContainerStarted","Data":"1cc0d2397e8835be72f1670cc38afcb4010d60869a22f58b59b7d9aef0ec825f"} Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.787816 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkgx"] Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.788221 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.788373 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.788594 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.788678 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.788726 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3180168d-9dad-441d-86bd-9d1801ef629f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.788930 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.789255 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.789287 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3180168d-9dad-441d-86bd-9d1801ef629f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.792982 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5b5k\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-kube-api-access-s5b5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.793157 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.793220 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: W1203 17:34:03.803689 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbd624460_116c_4630_b418_788f93ede485.slice/crio-8c9064e9250e611b615be754744ef9580e57225dac0c1ca338c3e28cbaece8d6 WatchSource:0}: Error finding container 8c9064e9250e611b615be754744ef9580e57225dac0c1ca338c3e28cbaece8d6: Status 404 returned error can't find the container with id 8c9064e9250e611b615be754744ef9580e57225dac0c1ca338c3e28cbaece8d6 Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895422 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895492 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895521 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3180168d-9dad-441d-86bd-9d1801ef629f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895548 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5b5k\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-kube-api-access-s5b5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895589 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895620 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895641 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895685 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895717 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895761 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895788 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3180168d-9dad-441d-86bd-9d1801ef629f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.895969 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.896256 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.896674 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.899073 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.899197 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.899456 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.899601 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.899672 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.903722 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3180168d-9dad-441d-86bd-9d1801ef629f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.907042 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.907417 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.908282 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.910335 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3180168d-9dad-441d-86bd-9d1801ef629f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.910608 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.913517 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.916535 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5b5k\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-kube-api-access-s5b5k\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.922184 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:03 crc kubenswrapper[4787]: I1203 17:34:03.936580 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.018214 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.020832 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.025679 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.026716 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.026922 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mpwfm" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.027048 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.027152 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.027263 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.027370 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.034995 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6rzx9" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.043348 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.067201 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.111370 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.111419 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.111468 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.111542 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-config-data\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.111599 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.111716 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d73396bd-2e1f-4c20-8327-610c5d783373-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.111795 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.111850 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5lg9b\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-kube-api-access-5lg9b\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.111904 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d73396bd-2e1f-4c20-8327-610c5d783373-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.111999 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.112067 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.215190 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.215251 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.215295 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.215330 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-config-data\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.215362 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.215404 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d73396bd-2e1f-4c20-8327-610c5d783373-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.215441 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.215474 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5lg9b\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-kube-api-access-5lg9b\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.215502 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d73396bd-2e1f-4c20-8327-610c5d783373-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.215542 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.215566 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.216233 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.216375 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.227908 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.230323 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.230971 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-config-data\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.231109 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.234638 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d73396bd-2e1f-4c20-8327-610c5d783373-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.237733 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.243626 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d73396bd-2e1f-4c20-8327-610c5d783373-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.257894 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5lg9b\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-kube-api-access-5lg9b\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.259007 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.295219 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.354241 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.791081 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" event={"ID":"bd624460-116c-4630-b418-788f93ede485","Type":"ContainerStarted","Data":"8c9064e9250e611b615be754744ef9580e57225dac0c1ca338c3e28cbaece8d6"} Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.798857 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:34:04 crc kubenswrapper[4787]: I1203 17:34:04.915763 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:34:05 crc kubenswrapper[4787]: W1203 17:34:05.016218 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd73396bd_2e1f_4c20_8327_610c5d783373.slice/crio-0cba0f8daa821c71af0478f2451d263670606b44b0ba3acb56917d287324d6aa WatchSource:0}: Error finding container 0cba0f8daa821c71af0478f2451d263670606b44b0ba3acb56917d287324d6aa: Status 404 returned error can't find the container with id 0cba0f8daa821c71af0478f2451d263670606b44b0ba3acb56917d287324d6aa Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.640958 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.643237 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.651803 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.652791 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.652936 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.653144 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-w7g9z" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.664722 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.666642 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.791958 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/02a278fe-fdca-423f-a302-59a08cf74ba4-config-data-default\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.792001 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.792034 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/02a278fe-fdca-423f-a302-59a08cf74ba4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.792080 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a278fe-fdca-423f-a302-59a08cf74ba4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.792098 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/02a278fe-fdca-423f-a302-59a08cf74ba4-kolla-config\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.792113 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02a278fe-fdca-423f-a302-59a08cf74ba4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.792153 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/02a278fe-fdca-423f-a302-59a08cf74ba4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.792170 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmjrc\" (UniqueName: \"kubernetes.io/projected/02a278fe-fdca-423f-a302-59a08cf74ba4-kube-api-access-wmjrc\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.896338 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/02a278fe-fdca-423f-a302-59a08cf74ba4-config-data-default\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.896408 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.896457 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/02a278fe-fdca-423f-a302-59a08cf74ba4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.896523 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a278fe-fdca-423f-a302-59a08cf74ba4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.896548 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/02a278fe-fdca-423f-a302-59a08cf74ba4-kolla-config\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.896566 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02a278fe-fdca-423f-a302-59a08cf74ba4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.896651 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/02a278fe-fdca-423f-a302-59a08cf74ba4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.896672 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmjrc\" (UniqueName: \"kubernetes.io/projected/02a278fe-fdca-423f-a302-59a08cf74ba4-kube-api-access-wmjrc\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.913911 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/02a278fe-fdca-423f-a302-59a08cf74ba4-config-data-default\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.920294 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.915009 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/02a278fe-fdca-423f-a302-59a08cf74ba4-kolla-config\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.926279 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/02a278fe-fdca-423f-a302-59a08cf74ba4-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.927354 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/02a278fe-fdca-423f-a302-59a08cf74ba4-config-data-generated\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.932667 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/02a278fe-fdca-423f-a302-59a08cf74ba4-operator-scripts\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.945789 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmjrc\" (UniqueName: \"kubernetes.io/projected/02a278fe-fdca-423f-a302-59a08cf74ba4-kube-api-access-wmjrc\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.957358 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d73396bd-2e1f-4c20-8327-610c5d783373","Type":"ContainerStarted","Data":"0cba0f8daa821c71af0478f2451d263670606b44b0ba3acb56917d287324d6aa"} Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.960319 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/02a278fe-fdca-423f-a302-59a08cf74ba4-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:05 crc kubenswrapper[4787]: I1203 17:34:05.960374 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3180168d-9dad-441d-86bd-9d1801ef629f","Type":"ContainerStarted","Data":"e562f65d2903616d598009a52a3d99823bf9bd894c41b5b25dfb0a9ca535e634"} Dec 03 17:34:06 crc kubenswrapper[4787]: I1203 17:34:06.004975 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"openstack-galera-0\" (UID: \"02a278fe-fdca-423f-a302-59a08cf74ba4\") " pod="openstack/openstack-galera-0" Dec 03 17:34:06 crc kubenswrapper[4787]: I1203 17:34:06.297450 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 03 17:34:06 crc kubenswrapper[4787]: I1203 17:34:06.936160 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.061262 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.062984 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.069880 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.081140 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.083197 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.083454 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-m646v" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.102092 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.207981 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.209952 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.225849 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.226413 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-bgnnl" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.226668 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.261431 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d207ad18-746f-4012-a399-2876e6d6bc10-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.261493 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d207ad18-746f-4012-a399-2876e6d6bc10-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.261531 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d207ad18-746f-4012-a399-2876e6d6bc10-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.261550 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.276390 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.278717 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d207ad18-746f-4012-a399-2876e6d6bc10-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.278816 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d207ad18-746f-4012-a399-2876e6d6bc10-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.278840 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d207ad18-746f-4012-a399-2876e6d6bc10-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.278891 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7g4mm\" (UniqueName: \"kubernetes.io/projected/d207ad18-746f-4012-a399-2876e6d6bc10-kube-api-access-7g4mm\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.386815 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2qqk5\" (UniqueName: \"kubernetes.io/projected/fe87a100-956b-468e-8aa3-164935a19f5f-kube-api-access-2qqk5\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.386861 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d207ad18-746f-4012-a399-2876e6d6bc10-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.386880 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe87a100-956b-468e-8aa3-164935a19f5f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.386898 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.386943 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d207ad18-746f-4012-a399-2876e6d6bc10-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.386971 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe87a100-956b-468e-8aa3-164935a19f5f-kolla-config\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.387033 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d207ad18-746f-4012-a399-2876e6d6bc10-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.387054 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe87a100-956b-468e-8aa3-164935a19f5f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.387073 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d207ad18-746f-4012-a399-2876e6d6bc10-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.387097 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7g4mm\" (UniqueName: \"kubernetes.io/projected/d207ad18-746f-4012-a399-2876e6d6bc10-kube-api-access-7g4mm\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.387129 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fe87a100-956b-468e-8aa3-164935a19f5f-config-data\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.387168 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d207ad18-746f-4012-a399-2876e6d6bc10-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.387195 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d207ad18-746f-4012-a399-2876e6d6bc10-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.388140 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/d207ad18-746f-4012-a399-2876e6d6bc10-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.388228 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/d207ad18-746f-4012-a399-2876e6d6bc10-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.388837 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/d207ad18-746f-4012-a399-2876e6d6bc10-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.389031 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.389195 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d207ad18-746f-4012-a399-2876e6d6bc10-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.418253 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d207ad18-746f-4012-a399-2876e6d6bc10-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.466197 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7g4mm\" (UniqueName: \"kubernetes.io/projected/d207ad18-746f-4012-a399-2876e6d6bc10-kube-api-access-7g4mm\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.484413 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.488814 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe87a100-956b-468e-8aa3-164935a19f5f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.488895 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fe87a100-956b-468e-8aa3-164935a19f5f-config-data\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.488957 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2qqk5\" (UniqueName: \"kubernetes.io/projected/fe87a100-956b-468e-8aa3-164935a19f5f-kube-api-access-2qqk5\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.488976 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe87a100-956b-468e-8aa3-164935a19f5f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.489047 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe87a100-956b-468e-8aa3-164935a19f5f-kolla-config\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.490150 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/fe87a100-956b-468e-8aa3-164935a19f5f-kolla-config\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.490888 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fe87a100-956b-468e-8aa3-164935a19f5f-config-data\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.492546 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe87a100-956b-468e-8aa3-164935a19f5f-combined-ca-bundle\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.492887 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe87a100-956b-468e-8aa3-164935a19f5f-memcached-tls-certs\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.500937 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/d207ad18-746f-4012-a399-2876e6d6bc10-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"d207ad18-746f-4012-a399-2876e6d6bc10\") " pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.517475 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2qqk5\" (UniqueName: \"kubernetes.io/projected/fe87a100-956b-468e-8aa3-164935a19f5f-kube-api-access-2qqk5\") pod \"memcached-0\" (UID: \"fe87a100-956b-468e-8aa3-164935a19f5f\") " pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.578240 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 03 17:34:07 crc kubenswrapper[4787]: I1203 17:34:07.690987 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:08 crc kubenswrapper[4787]: I1203 17:34:08.000186 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"02a278fe-fdca-423f-a302-59a08cf74ba4","Type":"ContainerStarted","Data":"4aa86c806928923264eee9432e4cc243aac9d403c79e78031d0bcd17907cac09"} Dec 03 17:34:08 crc kubenswrapper[4787]: I1203 17:34:08.564414 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 03 17:34:08 crc kubenswrapper[4787]: W1203 17:34:08.584384 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe87a100_956b_468e_8aa3_164935a19f5f.slice/crio-4b7ebd135e7a42d920669600cac841969d2d49e094ba567f48a6d8a581f5b886 WatchSource:0}: Error finding container 4b7ebd135e7a42d920669600cac841969d2d49e094ba567f48a6d8a581f5b886: Status 404 returned error can't find the container with id 4b7ebd135e7a42d920669600cac841969d2d49e094ba567f48a6d8a581f5b886 Dec 03 17:34:08 crc kubenswrapper[4787]: I1203 17:34:08.749843 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 03 17:34:08 crc kubenswrapper[4787]: W1203 17:34:08.762238 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd207ad18_746f_4012_a399_2876e6d6bc10.slice/crio-1cde414aa7af0d0ef86238da808bc644b4067693a5b8b48f82eabf533b241f67 WatchSource:0}: Error finding container 1cde414aa7af0d0ef86238da808bc644b4067693a5b8b48f82eabf533b241f67: Status 404 returned error can't find the container with id 1cde414aa7af0d0ef86238da808bc644b4067693a5b8b48f82eabf533b241f67 Dec 03 17:34:09 crc kubenswrapper[4787]: I1203 17:34:09.026923 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fe87a100-956b-468e-8aa3-164935a19f5f","Type":"ContainerStarted","Data":"4b7ebd135e7a42d920669600cac841969d2d49e094ba567f48a6d8a581f5b886"} Dec 03 17:34:09 crc kubenswrapper[4787]: I1203 17:34:09.035532 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d207ad18-746f-4012-a399-2876e6d6bc10","Type":"ContainerStarted","Data":"1cde414aa7af0d0ef86238da808bc644b4067693a5b8b48f82eabf533b241f67"} Dec 03 17:34:09 crc kubenswrapper[4787]: I1203 17:34:09.157196 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:34:09 crc kubenswrapper[4787]: I1203 17:34:09.158694 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:34:09 crc kubenswrapper[4787]: I1203 17:34:09.165414 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-6cb2r" Dec 03 17:34:09 crc kubenswrapper[4787]: I1203 17:34:09.192900 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:34:09 crc kubenswrapper[4787]: I1203 17:34:09.332996 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6977\" (UniqueName: \"kubernetes.io/projected/f5a1dad4-bf28-4c9a-987f-c66658f51ae7-kube-api-access-t6977\") pod \"kube-state-metrics-0\" (UID: \"f5a1dad4-bf28-4c9a-987f-c66658f51ae7\") " pod="openstack/kube-state-metrics-0" Dec 03 17:34:09 crc kubenswrapper[4787]: I1203 17:34:09.434689 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6977\" (UniqueName: \"kubernetes.io/projected/f5a1dad4-bf28-4c9a-987f-c66658f51ae7-kube-api-access-t6977\") pod \"kube-state-metrics-0\" (UID: \"f5a1dad4-bf28-4c9a-987f-c66658f51ae7\") " pod="openstack/kube-state-metrics-0" Dec 03 17:34:09 crc kubenswrapper[4787]: I1203 17:34:09.485058 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6977\" (UniqueName: \"kubernetes.io/projected/f5a1dad4-bf28-4c9a-987f-c66658f51ae7-kube-api-access-t6977\") pod \"kube-state-metrics-0\" (UID: \"f5a1dad4-bf28-4c9a-987f-c66658f51ae7\") " pod="openstack/kube-state-metrics-0" Dec 03 17:34:09 crc kubenswrapper[4787]: I1203 17:34:09.493217 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.045237 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr"] Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.046287 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.064351 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards-sa-dockercfg-b7vs6" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.064730 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-ui-dashboards" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.098182 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr"] Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.164250 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c72c745a-9a1c-4374-b1a7-97279603edf7-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-qwtfr\" (UID: \"c72c745a-9a1c-4374-b1a7-97279603edf7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.164386 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k8k8s\" (UniqueName: \"kubernetes.io/projected/c72c745a-9a1c-4374-b1a7-97279603edf7-kube-api-access-k8k8s\") pod \"observability-ui-dashboards-7d5fb4cbfb-qwtfr\" (UID: \"c72c745a-9a1c-4374-b1a7-97279603edf7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.266976 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c72c745a-9a1c-4374-b1a7-97279603edf7-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-qwtfr\" (UID: \"c72c745a-9a1c-4374-b1a7-97279603edf7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.267081 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k8k8s\" (UniqueName: \"kubernetes.io/projected/c72c745a-9a1c-4374-b1a7-97279603edf7-kube-api-access-k8k8s\") pod \"observability-ui-dashboards-7d5fb4cbfb-qwtfr\" (UID: \"c72c745a-9a1c-4374-b1a7-97279603edf7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" Dec 03 17:34:10 crc kubenswrapper[4787]: E1203 17:34:10.267190 4787 secret.go:188] Couldn't get secret openshift-operators/observability-ui-dashboards: secret "observability-ui-dashboards" not found Dec 03 17:34:10 crc kubenswrapper[4787]: E1203 17:34:10.267258 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c72c745a-9a1c-4374-b1a7-97279603edf7-serving-cert podName:c72c745a-9a1c-4374-b1a7-97279603edf7 nodeName:}" failed. No retries permitted until 2025-12-03 17:34:10.767241012 +0000 UTC m=+1267.584711971 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/c72c745a-9a1c-4374-b1a7-97279603edf7-serving-cert") pod "observability-ui-dashboards-7d5fb4cbfb-qwtfr" (UID: "c72c745a-9a1c-4374-b1a7-97279603edf7") : secret "observability-ui-dashboards" not found Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.293373 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k8k8s\" (UniqueName: \"kubernetes.io/projected/c72c745a-9a1c-4374-b1a7-97279603edf7-kube-api-access-k8k8s\") pod \"observability-ui-dashboards-7d5fb4cbfb-qwtfr\" (UID: \"c72c745a-9a1c-4374-b1a7-97279603edf7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.417459 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-6f77dc657b-v8bks"] Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.422909 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.476146 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f77dc657b-v8bks"] Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.517376 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.521697 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.525388 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.528936 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.531324 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.531628 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.531867 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-4g9vz" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.532145 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.542203 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.574828 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-console-oauth-config\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.575137 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-service-ca\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.575169 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-console-config\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.575266 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-oauth-serving-cert\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.575285 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnvhl\" (UniqueName: \"kubernetes.io/projected/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-kube-api-access-pnvhl\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.575302 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-console-serving-cert\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.575330 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-trusted-ca-bundle\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681393 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-console-oauth-config\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681433 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-service-ca\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681464 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681485 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-console-config\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681545 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3b68d021-5db8-4a87-92f2-e827bbdd2255-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681572 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681625 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-config\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681643 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n62gz\" (UniqueName: \"kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-kube-api-access-n62gz\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681662 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681699 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-oauth-serving-cert\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681715 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnvhl\" (UniqueName: \"kubernetes.io/projected/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-kube-api-access-pnvhl\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681733 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-console-serving-cert\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681759 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-trusted-ca-bundle\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681776 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3b68d021-5db8-4a87-92f2-e827bbdd2255-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.681806 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.683252 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-oauth-serving-cert\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.683655 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-console-config\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.684343 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-service-ca\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.684500 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-trusted-ca-bundle\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.690552 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-console-serving-cert\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.697834 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-console-oauth-config\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.717729 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnvhl\" (UniqueName: \"kubernetes.io/projected/ed1b7b6c-4e84-45a5-8915-1ec509e98e78-kube-api-access-pnvhl\") pod \"console-6f77dc657b-v8bks\" (UID: \"ed1b7b6c-4e84-45a5-8915-1ec509e98e78\") " pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.783224 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-config\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.783266 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.783286 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n62gz\" (UniqueName: \"kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-kube-api-access-n62gz\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.783387 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3b68d021-5db8-4a87-92f2-e827bbdd2255-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.783413 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c72c745a-9a1c-4374-b1a7-97279603edf7-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-qwtfr\" (UID: \"c72c745a-9a1c-4374-b1a7-97279603edf7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.783429 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.783458 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.783498 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3b68d021-5db8-4a87-92f2-e827bbdd2255-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.783527 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.784304 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.788997 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-config\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.789501 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.789767 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.790864 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.791846 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3b68d021-5db8-4a87-92f2-e827bbdd2255-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.794265 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c72c745a-9a1c-4374-b1a7-97279603edf7-serving-cert\") pod \"observability-ui-dashboards-7d5fb4cbfb-qwtfr\" (UID: \"c72c745a-9a1c-4374-b1a7-97279603edf7\") " pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.796462 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3b68d021-5db8-4a87-92f2-e827bbdd2255-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.796510 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.799695 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n62gz\" (UniqueName: \"kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-kube-api-access-n62gz\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.819802 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.875274 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 17:34:10 crc kubenswrapper[4787]: I1203 17:34:10.995214 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.229495 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-prdc5"] Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.231218 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.235801 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-6bc74" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.235974 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.237170 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.253201 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-prdc5"] Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.272931 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-mc6r6"] Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.276433 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.298607 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mc6r6"] Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.335715 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2db4ba58-5d6c-42ff-b386-7765170d9113-scripts\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.335815 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2db4ba58-5d6c-42ff-b386-7765170d9113-var-log-ovn\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.335852 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2db4ba58-5d6c-42ff-b386-7765170d9113-var-run\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.335988 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db4ba58-5d6c-42ff-b386-7765170d9113-ovn-controller-tls-certs\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.336038 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmpjn\" (UniqueName: \"kubernetes.io/projected/2db4ba58-5d6c-42ff-b386-7765170d9113-kube-api-access-qmpjn\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.336311 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2db4ba58-5d6c-42ff-b386-7765170d9113-var-run-ovn\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.336357 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db4ba58-5d6c-42ff-b386-7765170d9113-combined-ca-bundle\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.438571 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db4ba58-5d6c-42ff-b386-7765170d9113-ovn-controller-tls-certs\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.438891 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmpjn\" (UniqueName: \"kubernetes.io/projected/2db4ba58-5d6c-42ff-b386-7765170d9113-kube-api-access-qmpjn\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.438916 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-var-lib\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.438955 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2db4ba58-5d6c-42ff-b386-7765170d9113-var-run-ovn\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.438978 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db4ba58-5d6c-42ff-b386-7765170d9113-combined-ca-bundle\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.438999 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-var-run\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.439042 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2db4ba58-5d6c-42ff-b386-7765170d9113-scripts\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.439067 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cab2a5b3-1157-45ab-ad4b-828a1638fd78-scripts\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.439085 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-var-log\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.439102 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-etc-ovs\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.439127 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2db4ba58-5d6c-42ff-b386-7765170d9113-var-log-ovn\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.439148 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2db4ba58-5d6c-42ff-b386-7765170d9113-var-run\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.439206 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwn77\" (UniqueName: \"kubernetes.io/projected/cab2a5b3-1157-45ab-ad4b-828a1638fd78-kube-api-access-kwn77\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.442717 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/2db4ba58-5d6c-42ff-b386-7765170d9113-scripts\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.443425 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/2db4ba58-5d6c-42ff-b386-7765170d9113-var-run-ovn\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.444550 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/2db4ba58-5d6c-42ff-b386-7765170d9113-var-run\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.444531 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/2db4ba58-5d6c-42ff-b386-7765170d9113-var-log-ovn\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.454110 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2db4ba58-5d6c-42ff-b386-7765170d9113-combined-ca-bundle\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.458545 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmpjn\" (UniqueName: \"kubernetes.io/projected/2db4ba58-5d6c-42ff-b386-7765170d9113-kube-api-access-qmpjn\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.465238 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/2db4ba58-5d6c-42ff-b386-7765170d9113-ovn-controller-tls-certs\") pod \"ovn-controller-prdc5\" (UID: \"2db4ba58-5d6c-42ff-b386-7765170d9113\") " pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.540260 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwn77\" (UniqueName: \"kubernetes.io/projected/cab2a5b3-1157-45ab-ad4b-828a1638fd78-kube-api-access-kwn77\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.540367 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-var-lib\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.540406 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-var-run\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.540433 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-var-log\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.540447 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cab2a5b3-1157-45ab-ad4b-828a1638fd78-scripts\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.540465 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-etc-ovs\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.540671 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-var-lib\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.540681 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-var-log\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.540728 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-etc-ovs\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.540737 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cab2a5b3-1157-45ab-ad4b-828a1638fd78-var-run\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.543079 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cab2a5b3-1157-45ab-ad4b-828a1638fd78-scripts\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.559856 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwn77\" (UniqueName: \"kubernetes.io/projected/cab2a5b3-1157-45ab-ad4b-828a1638fd78-kube-api-access-kwn77\") pod \"ovn-controller-ovs-mc6r6\" (UID: \"cab2a5b3-1157-45ab-ad4b-828a1638fd78\") " pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.575711 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-prdc5" Dec 03 17:34:12 crc kubenswrapper[4787]: I1203 17:34:12.605221 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.454775 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.457058 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.460722 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.460961 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.461160 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.461349 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-5lld7" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.461509 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.481382 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.625800 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd30e7ed-2464-4857-a563-b02f14871d54-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.625950 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd30e7ed-2464-4857-a563-b02f14871d54-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.625990 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd30e7ed-2464-4857-a563-b02f14871d54-config\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.626094 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2mqn\" (UniqueName: \"kubernetes.io/projected/cd30e7ed-2464-4857-a563-b02f14871d54-kube-api-access-c2mqn\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.626110 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd30e7ed-2464-4857-a563-b02f14871d54-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.626137 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd30e7ed-2464-4857-a563-b02f14871d54-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.626178 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.626239 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cd30e7ed-2464-4857-a563-b02f14871d54-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.728072 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2mqn\" (UniqueName: \"kubernetes.io/projected/cd30e7ed-2464-4857-a563-b02f14871d54-kube-api-access-c2mqn\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.728127 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd30e7ed-2464-4857-a563-b02f14871d54-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.728151 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd30e7ed-2464-4857-a563-b02f14871d54-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.728181 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.728218 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cd30e7ed-2464-4857-a563-b02f14871d54-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.728250 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd30e7ed-2464-4857-a563-b02f14871d54-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.728302 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd30e7ed-2464-4857-a563-b02f14871d54-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.728328 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd30e7ed-2464-4857-a563-b02f14871d54-config\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.728687 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.728870 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/cd30e7ed-2464-4857-a563-b02f14871d54-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.729591 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cd30e7ed-2464-4857-a563-b02f14871d54-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.730468 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd30e7ed-2464-4857-a563-b02f14871d54-config\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.734787 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd30e7ed-2464-4857-a563-b02f14871d54-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.739164 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd30e7ed-2464-4857-a563-b02f14871d54-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.739845 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd30e7ed-2464-4857-a563-b02f14871d54-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.747647 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2mqn\" (UniqueName: \"kubernetes.io/projected/cd30e7ed-2464-4857-a563-b02f14871d54-kube-api-access-c2mqn\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.757096 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"cd30e7ed-2464-4857-a563-b02f14871d54\") " pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.781701 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.801471 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.809586 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.813851 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-rdgtb" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.814213 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.815908 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.816128 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.820839 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.931666 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.931975 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b71b702-ef57-41ac-8ada-1fe782d5092a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.933254 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b71b702-ef57-41ac-8ada-1fe782d5092a-config\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.933990 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjzp6\" (UniqueName: \"kubernetes.io/projected/6b71b702-ef57-41ac-8ada-1fe782d5092a-kube-api-access-tjzp6\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.934166 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b71b702-ef57-41ac-8ada-1fe782d5092a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.934360 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6b71b702-ef57-41ac-8ada-1fe782d5092a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.934502 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b71b702-ef57-41ac-8ada-1fe782d5092a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:15 crc kubenswrapper[4787]: I1203 17:34:15.934615 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b71b702-ef57-41ac-8ada-1fe782d5092a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.036332 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b71b702-ef57-41ac-8ada-1fe782d5092a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.036458 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.036495 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b71b702-ef57-41ac-8ada-1fe782d5092a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.036550 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b71b702-ef57-41ac-8ada-1fe782d5092a-config\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.036647 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjzp6\" (UniqueName: \"kubernetes.io/projected/6b71b702-ef57-41ac-8ada-1fe782d5092a-kube-api-access-tjzp6\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.036696 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b71b702-ef57-41ac-8ada-1fe782d5092a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.036730 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6b71b702-ef57-41ac-8ada-1fe782d5092a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.036758 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b71b702-ef57-41ac-8ada-1fe782d5092a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.037421 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.038275 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6b71b702-ef57-41ac-8ada-1fe782d5092a-config\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.038304 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b71b702-ef57-41ac-8ada-1fe782d5092a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.038739 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/6b71b702-ef57-41ac-8ada-1fe782d5092a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.041272 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b71b702-ef57-41ac-8ada-1fe782d5092a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.042139 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b71b702-ef57-41ac-8ada-1fe782d5092a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.042611 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b71b702-ef57-41ac-8ada-1fe782d5092a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.054715 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjzp6\" (UniqueName: \"kubernetes.io/projected/6b71b702-ef57-41ac-8ada-1fe782d5092a-kube-api-access-tjzp6\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.064088 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"6b71b702-ef57-41ac-8ada-1fe782d5092a\") " pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:16 crc kubenswrapper[4787]: I1203 17:34:16.150718 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:18 crc kubenswrapper[4787]: I1203 17:34:18.989912 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:34:18 crc kubenswrapper[4787]: I1203 17:34:18.990340 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:34:18 crc kubenswrapper[4787]: I1203 17:34:18.990396 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:34:18 crc kubenswrapper[4787]: I1203 17:34:18.991246 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"93e6b55e4dd0c75878fc7a5cbe5a33dad37bb97aa5cbfc657a6c1fd769073445"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:34:18 crc kubenswrapper[4787]: I1203 17:34:18.991308 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://93e6b55e4dd0c75878fc7a5cbe5a33dad37bb97aa5cbfc657a6c1fd769073445" gracePeriod=600 Dec 03 17:34:20 crc kubenswrapper[4787]: I1203 17:34:20.174378 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="93e6b55e4dd0c75878fc7a5cbe5a33dad37bb97aa5cbfc657a6c1fd769073445" exitCode=0 Dec 03 17:34:20 crc kubenswrapper[4787]: I1203 17:34:20.174607 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"93e6b55e4dd0c75878fc7a5cbe5a33dad37bb97aa5cbfc657a6c1fd769073445"} Dec 03 17:34:20 crc kubenswrapper[4787]: I1203 17:34:20.174770 4787 scope.go:117] "RemoveContainer" containerID="864b789057addccde661186743b5963a0f8e318c2a7a639d2b92c72f83cc12f6" Dec 03 17:34:25 crc kubenswrapper[4787]: E1203 17:34:25.262709 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 03 17:34:25 crc kubenswrapper[4787]: E1203 17:34:25.263431 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5lg9b,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-server-0_openstack(d73396bd-2e1f-4c20-8327-610c5d783373): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:34:25 crc kubenswrapper[4787]: E1203 17:34:25.264737 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-server-0" podUID="d73396bd-2e1f-4c20-8327-610c5d783373" Dec 03 17:34:25 crc kubenswrapper[4787]: E1203 17:34:25.289551 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified" Dec 03 17:34:25 crc kubenswrapper[4787]: E1203 17:34:25.289980 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:setup-container,Image:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,Command:[sh -c cp /tmp/erlang-cookie-secret/.erlang.cookie /var/lib/rabbitmq/.erlang.cookie && chmod 600 /var/lib/rabbitmq/.erlang.cookie ; cp /tmp/rabbitmq-plugins/enabled_plugins /operator/enabled_plugins ; echo '[default]' > /var/lib/rabbitmq/.rabbitmqadmin.conf && sed -e 's/default_user/username/' -e 's/default_pass/password/' /tmp/default_user.conf >> /var/lib/rabbitmq/.rabbitmqadmin.conf && chmod 600 /var/lib/rabbitmq/.rabbitmqadmin.conf ; sleep 30],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Requests:ResourceList{cpu: {{20 -3} {} 20m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:plugins-conf,ReadOnly:false,MountPath:/tmp/rabbitmq-plugins/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-erlang-cookie,ReadOnly:false,MountPath:/var/lib/rabbitmq/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:erlang-cookie-secret,ReadOnly:false,MountPath:/tmp/erlang-cookie-secret/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-plugins,ReadOnly:false,MountPath:/operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:persistence,ReadOnly:false,MountPath:/var/lib/rabbitmq/mnesia/,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:rabbitmq-confd,ReadOnly:false,MountPath:/tmp/default_user.conf,SubPath:default_user.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-s5b5k,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cell1-server-0_openstack(3180168d-9dad-441d-86bd-9d1801ef629f): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:34:25 crc kubenswrapper[4787]: E1203 17:34:25.291820 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/rabbitmq-cell1-server-0" podUID="3180168d-9dad-441d-86bd-9d1801ef629f" Dec 03 17:34:26 crc kubenswrapper[4787]: E1203 17:34:26.221899 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-server-0" podUID="d73396bd-2e1f-4c20-8327-610c5d783373" Dec 03 17:34:26 crc kubenswrapper[4787]: E1203 17:34:26.222200 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"setup-container\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified\\\"\"" pod="openstack/rabbitmq-cell1-server-0" podUID="3180168d-9dad-441d-86bd-9d1801ef629f" Dec 03 17:34:30 crc kubenswrapper[4787]: I1203 17:34:30.794618 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 17:34:30 crc kubenswrapper[4787]: I1203 17:34:30.856089 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-6f77dc657b-v8bks"] Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.567820 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.568306 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nr6rl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-qmdcv_openstack(e231ba27-59a3-48a6-b716-cb1a18a07602): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.569889 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" podUID="e231ba27-59a3-48a6-b716-cb1a18a07602" Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.586892 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.587466 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-495zj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-kd2cd_openstack(85be111d-8722-48ec-8f70-6c8c53c0fcd8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.590488 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" podUID="85be111d-8722-48ec-8f70-6c8c53c0fcd8" Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.684514 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.684708 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bl59w,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-57d769cc4f-bmkgx_openstack(bd624460-116c-4630-b418-788f93ede485): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.684871 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.684980 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nw9ms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-dxzmh_openstack(5dc53172-a8d8-44f8-9969-d5ed647dcd71): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.690211 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" podUID="5dc53172-a8d8-44f8-9969-d5ed647dcd71" Dec 03 17:34:31 crc kubenswrapper[4787]: E1203 17:34:31.690328 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" podUID="bd624460-116c-4630-b418-788f93ede485" Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.303514 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"8530351cb9d4d0e6acbfd6734059c845290cade8ecc29a25ad6ed13ae5d5db93"} Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.305844 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"fe87a100-956b-468e-8aa3-164935a19f5f","Type":"ContainerStarted","Data":"c8de0e4e1819d519425dd6e2000bb03a297dc66797aee18452f00299586e27d5"} Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.306353 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.308293 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f77dc657b-v8bks" event={"ID":"ed1b7b6c-4e84-45a5-8915-1ec509e98e78","Type":"ContainerStarted","Data":"b8a95614a6cf2b084e1b8f2feb245c0327843ac81c01a1c8ecdbd6633058d487"} Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.308320 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-6f77dc657b-v8bks" event={"ID":"ed1b7b6c-4e84-45a5-8915-1ec509e98e78","Type":"ContainerStarted","Data":"6991fde7b0bfe6b713df365683d3c354dc5ca1685d861f0c60a5cadb615b3bc2"} Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.310438 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3b68d021-5db8-4a87-92f2-e827bbdd2255","Type":"ContainerStarted","Data":"e2aa77b4845468582d3647e9a42860dcb980d9e2227bdd8179a216ef2686ac89"} Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.312159 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d207ad18-746f-4012-a399-2876e6d6bc10","Type":"ContainerStarted","Data":"3720cceff0c8c190bfb53ffb58251c119cf87dcdbf085507bb319c852d044e6b"} Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.335944 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"02a278fe-fdca-423f-a302-59a08cf74ba4","Type":"ContainerStarted","Data":"4ff02e05f143829b60899551a2ce935473afb5c289c3324a07e439e6bf58cffb"} Dec 03 17:34:32 crc kubenswrapper[4787]: E1203 17:34:32.338725 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" podUID="bd624460-116c-4630-b418-788f93ede485" Dec 03 17:34:32 crc kubenswrapper[4787]: E1203 17:34:32.339076 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified\\\"\"" pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" podUID="5dc53172-a8d8-44f8-9969-d5ed647dcd71" Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.366435 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-prdc5"] Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.419330 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr"] Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.443640 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.476872 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-6f77dc657b-v8bks" podStartSLOduration=22.476856057 podStartE2EDuration="22.476856057s" podCreationTimestamp="2025-12-03 17:34:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:34:32.359217476 +0000 UTC m=+1289.176688455" watchObservedRunningTime="2025-12-03 17:34:32.476856057 +0000 UTC m=+1289.294327016" Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.501660 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=2.513427847 podStartE2EDuration="25.501642563s" podCreationTimestamp="2025-12-03 17:34:07 +0000 UTC" firstStartedPulling="2025-12-03 17:34:08.590118386 +0000 UTC m=+1265.407589345" lastFinishedPulling="2025-12-03 17:34:31.578333102 +0000 UTC m=+1288.395804061" observedRunningTime="2025-12-03 17:34:32.441175903 +0000 UTC m=+1289.258646872" watchObservedRunningTime="2025-12-03 17:34:32.501642563 +0000 UTC m=+1289.319113522" Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.764720 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 03 17:34:32 crc kubenswrapper[4787]: I1203 17:34:32.928630 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.037257 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.055414 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85be111d-8722-48ec-8f70-6c8c53c0fcd8-config\") pod \"85be111d-8722-48ec-8f70-6c8c53c0fcd8\" (UID: \"85be111d-8722-48ec-8f70-6c8c53c0fcd8\") " Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.055736 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-495zj\" (UniqueName: \"kubernetes.io/projected/85be111d-8722-48ec-8f70-6c8c53c0fcd8-kube-api-access-495zj\") pod \"85be111d-8722-48ec-8f70-6c8c53c0fcd8\" (UID: \"85be111d-8722-48ec-8f70-6c8c53c0fcd8\") " Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.059893 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/85be111d-8722-48ec-8f70-6c8c53c0fcd8-config" (OuterVolumeSpecName: "config") pod "85be111d-8722-48ec-8f70-6c8c53c0fcd8" (UID: "85be111d-8722-48ec-8f70-6c8c53c0fcd8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.063277 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85be111d-8722-48ec-8f70-6c8c53c0fcd8-kube-api-access-495zj" (OuterVolumeSpecName: "kube-api-access-495zj") pod "85be111d-8722-48ec-8f70-6c8c53c0fcd8" (UID: "85be111d-8722-48ec-8f70-6c8c53c0fcd8"). InnerVolumeSpecName "kube-api-access-495zj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.156804 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nr6rl\" (UniqueName: \"kubernetes.io/projected/e231ba27-59a3-48a6-b716-cb1a18a07602-kube-api-access-nr6rl\") pod \"e231ba27-59a3-48a6-b716-cb1a18a07602\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.156861 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-config\") pod \"e231ba27-59a3-48a6-b716-cb1a18a07602\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.156954 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-dns-svc\") pod \"e231ba27-59a3-48a6-b716-cb1a18a07602\" (UID: \"e231ba27-59a3-48a6-b716-cb1a18a07602\") " Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.157451 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/85be111d-8722-48ec-8f70-6c8c53c0fcd8-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.157467 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-495zj\" (UniqueName: \"kubernetes.io/projected/85be111d-8722-48ec-8f70-6c8c53c0fcd8-kube-api-access-495zj\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.157845 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "e231ba27-59a3-48a6-b716-cb1a18a07602" (UID: "e231ba27-59a3-48a6-b716-cb1a18a07602"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.158625 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-config" (OuterVolumeSpecName: "config") pod "e231ba27-59a3-48a6-b716-cb1a18a07602" (UID: "e231ba27-59a3-48a6-b716-cb1a18a07602"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.161429 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e231ba27-59a3-48a6-b716-cb1a18a07602-kube-api-access-nr6rl" (OuterVolumeSpecName: "kube-api-access-nr6rl") pod "e231ba27-59a3-48a6-b716-cb1a18a07602" (UID: "e231ba27-59a3-48a6-b716-cb1a18a07602"). InnerVolumeSpecName "kube-api-access-nr6rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.259349 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nr6rl\" (UniqueName: \"kubernetes.io/projected/e231ba27-59a3-48a6-b716-cb1a18a07602-kube-api-access-nr6rl\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.259724 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.259736 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/e231ba27-59a3-48a6-b716-cb1a18a07602-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.347372 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" event={"ID":"e231ba27-59a3-48a6-b716-cb1a18a07602","Type":"ContainerDied","Data":"16713232cca87da1c8982a1d9811f2a676dd1ff00933d9807f0da473b86805a8"} Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.347471 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-qmdcv" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.350237 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" event={"ID":"c72c745a-9a1c-4374-b1a7-97279603edf7","Type":"ContainerStarted","Data":"54b5f8273cb808ed9f32843070922939b3ab175f1b64cb59836efe362960ff38"} Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.351793 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f5a1dad4-bf28-4c9a-987f-c66658f51ae7","Type":"ContainerStarted","Data":"7f55d21b4081d87824f14b26d6a717db9b12671e2b608b8e71255341c724f08e"} Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.353479 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6b71b702-ef57-41ac-8ada-1fe782d5092a","Type":"ContainerStarted","Data":"31d2543760840062efbc9b35084cc4e8efcb53b3a201c0a4abe78b738221d514"} Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.354847 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" event={"ID":"85be111d-8722-48ec-8f70-6c8c53c0fcd8","Type":"ContainerDied","Data":"93790da9084107b7a563c441b2ab0263c59c9bf1d65d75a601d1b65f3ced224e"} Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.354926 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-kd2cd" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.358719 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-prdc5" event={"ID":"2db4ba58-5d6c-42ff-b386-7765170d9113","Type":"ContainerStarted","Data":"0c8b7a57e66d89f4880ecc7d392699815fe080d7cb631224aff36e58b49ea275"} Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.430438 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qmdcv"] Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.448075 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-qmdcv"] Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.488916 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd2cd"] Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.523702 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-kd2cd"] Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.643270 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.810691 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85be111d-8722-48ec-8f70-6c8c53c0fcd8" path="/var/lib/kubelet/pods/85be111d-8722-48ec-8f70-6c8c53c0fcd8/volumes" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.811186 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e231ba27-59a3-48a6-b716-cb1a18a07602" path="/var/lib/kubelet/pods/e231ba27-59a3-48a6-b716-cb1a18a07602/volumes" Dec 03 17:34:33 crc kubenswrapper[4787]: I1203 17:34:33.811906 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-mc6r6"] Dec 03 17:34:34 crc kubenswrapper[4787]: I1203 17:34:34.373358 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mc6r6" event={"ID":"cab2a5b3-1157-45ab-ad4b-828a1638fd78","Type":"ContainerStarted","Data":"d5b5b8f07b80aeb975b9e775c8c9201c9920a9a192b94ef65eeff3053a3663aa"} Dec 03 17:34:35 crc kubenswrapper[4787]: I1203 17:34:35.384031 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cd30e7ed-2464-4857-a563-b02f14871d54","Type":"ContainerStarted","Data":"ba80dba7d6ce6fac60ade18a899dcb0625789ba7b14b02e2a0ef72b96f7d6508"} Dec 03 17:34:36 crc kubenswrapper[4787]: I1203 17:34:36.397406 4787 generic.go:334] "Generic (PLEG): container finished" podID="d207ad18-746f-4012-a399-2876e6d6bc10" containerID="3720cceff0c8c190bfb53ffb58251c119cf87dcdbf085507bb319c852d044e6b" exitCode=0 Dec 03 17:34:36 crc kubenswrapper[4787]: I1203 17:34:36.397473 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d207ad18-746f-4012-a399-2876e6d6bc10","Type":"ContainerDied","Data":"3720cceff0c8c190bfb53ffb58251c119cf87dcdbf085507bb319c852d044e6b"} Dec 03 17:34:36 crc kubenswrapper[4787]: I1203 17:34:36.403583 4787 generic.go:334] "Generic (PLEG): container finished" podID="02a278fe-fdca-423f-a302-59a08cf74ba4" containerID="4ff02e05f143829b60899551a2ce935473afb5c289c3324a07e439e6bf58cffb" exitCode=0 Dec 03 17:34:36 crc kubenswrapper[4787]: I1203 17:34:36.403627 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"02a278fe-fdca-423f-a302-59a08cf74ba4","Type":"ContainerDied","Data":"4ff02e05f143829b60899551a2ce935473afb5c289c3324a07e439e6bf58cffb"} Dec 03 17:34:37 crc kubenswrapper[4787]: I1203 17:34:37.581069 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.479687 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6b71b702-ef57-41ac-8ada-1fe782d5092a","Type":"ContainerStarted","Data":"5a7e65eceb7fb9c1b997ce26ed681af1040a783491c0608ae6eb3ca6514df30d"} Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.497992 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cd30e7ed-2464-4857-a563-b02f14871d54","Type":"ContainerStarted","Data":"a7d8ede4dba651b569d4499a65a1687d055ca3b1eca507ddd5f8a0a6635a30a0"} Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.518931 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mc6r6" event={"ID":"cab2a5b3-1157-45ab-ad4b-828a1638fd78","Type":"ContainerStarted","Data":"9fdfb1e0585fb00676e76659165c5315e8cf47e0535498c0f6c2ba0a7b67a3fe"} Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.529796 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"d207ad18-746f-4012-a399-2876e6d6bc10","Type":"ContainerStarted","Data":"64404ca43d96506e93a25a33ba33d5ef488710190a70f26aa881f101a66a3aa8"} Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.545506 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkgx"] Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.545724 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-prdc5" event={"ID":"2db4ba58-5d6c-42ff-b386-7765170d9113","Type":"ContainerStarted","Data":"5131ce9ba7528f2c5147d357982c5c4ea5d456d91693bcad184adfd460f95943"} Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.545793 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-prdc5" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.549251 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"02a278fe-fdca-423f-a302-59a08cf74ba4","Type":"ContainerStarted","Data":"3087cfda957395c2fb7d58f61ceeee9623a9ed2b477591ab8716d4fdcec60531"} Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.579672 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-99x2s"] Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.588885 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" event={"ID":"c72c745a-9a1c-4374-b1a7-97279603edf7","Type":"ContainerStarted","Data":"ec8a2424a1b3ac3db28eb30692869b7686e45fb7095596a8cb8a2a69eaa1f473"} Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.589080 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.593804 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f5a1dad4-bf28-4c9a-987f-c66658f51ae7","Type":"ContainerStarted","Data":"00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8"} Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.595175 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.608079 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-99x2s"] Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.630268 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9nzmp\" (UniqueName: \"kubernetes.io/projected/f8dd3942-d40c-4699-8603-74f41dd659be-kube-api-access-9nzmp\") pod \"dnsmasq-dns-7cb5889db5-99x2s\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.630848 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-config\") pod \"dnsmasq-dns-7cb5889db5-99x2s\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.631100 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-99x2s\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.692713 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-prdc5" podStartSLOduration=21.63241212 podStartE2EDuration="27.692693643s" podCreationTimestamp="2025-12-03 17:34:12 +0000 UTC" firstStartedPulling="2025-12-03 17:34:32.401164925 +0000 UTC m=+1289.218635884" lastFinishedPulling="2025-12-03 17:34:38.461446448 +0000 UTC m=+1295.278917407" observedRunningTime="2025-12-03 17:34:39.620976136 +0000 UTC m=+1296.438447115" watchObservedRunningTime="2025-12-03 17:34:39.692693643 +0000 UTC m=+1296.510164602" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.699429 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=12.363751865 podStartE2EDuration="35.699412551s" podCreationTimestamp="2025-12-03 17:34:04 +0000 UTC" firstStartedPulling="2025-12-03 17:34:06.992001508 +0000 UTC m=+1263.809472467" lastFinishedPulling="2025-12-03 17:34:30.327662174 +0000 UTC m=+1287.145133153" observedRunningTime="2025-12-03 17:34:39.645160046 +0000 UTC m=+1296.462631015" watchObservedRunningTime="2025-12-03 17:34:39.699412551 +0000 UTC m=+1296.516883510" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.708490 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=10.895835047 podStartE2EDuration="33.70846466s" podCreationTimestamp="2025-12-03 17:34:06 +0000 UTC" firstStartedPulling="2025-12-03 17:34:08.767819185 +0000 UTC m=+1265.585290134" lastFinishedPulling="2025-12-03 17:34:31.580448788 +0000 UTC m=+1288.397919747" observedRunningTime="2025-12-03 17:34:39.67027103 +0000 UTC m=+1296.487741999" watchObservedRunningTime="2025-12-03 17:34:39.70846466 +0000 UTC m=+1296.525935619" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.733312 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9nzmp\" (UniqueName: \"kubernetes.io/projected/f8dd3942-d40c-4699-8603-74f41dd659be-kube-api-access-9nzmp\") pod \"dnsmasq-dns-7cb5889db5-99x2s\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.734043 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-config\") pod \"dnsmasq-dns-7cb5889db5-99x2s\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.734332 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-99x2s\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.736080 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-dns-svc\") pod \"dnsmasq-dns-7cb5889db5-99x2s\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.737578 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-config\") pod \"dnsmasq-dns-7cb5889db5-99x2s\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.749869 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=24.665003232 podStartE2EDuration="30.749846825s" podCreationTimestamp="2025-12-03 17:34:09 +0000 UTC" firstStartedPulling="2025-12-03 17:34:32.474361131 +0000 UTC m=+1289.291832090" lastFinishedPulling="2025-12-03 17:34:38.559204724 +0000 UTC m=+1295.376675683" observedRunningTime="2025-12-03 17:34:39.711026848 +0000 UTC m=+1296.528497817" watchObservedRunningTime="2025-12-03 17:34:39.749846825 +0000 UTC m=+1296.567317784" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.765687 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-ui-dashboards-7d5fb4cbfb-qwtfr" podStartSLOduration=24.166799443 podStartE2EDuration="29.765668443s" podCreationTimestamp="2025-12-03 17:34:10 +0000 UTC" firstStartedPulling="2025-12-03 17:34:32.481421658 +0000 UTC m=+1289.298892617" lastFinishedPulling="2025-12-03 17:34:38.080290658 +0000 UTC m=+1294.897761617" observedRunningTime="2025-12-03 17:34:39.731871559 +0000 UTC m=+1296.549342538" watchObservedRunningTime="2025-12-03 17:34:39.765668443 +0000 UTC m=+1296.583139402" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.779442 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9nzmp\" (UniqueName: \"kubernetes.io/projected/f8dd3942-d40c-4699-8603-74f41dd659be-kube-api-access-9nzmp\") pod \"dnsmasq-dns-7cb5889db5-99x2s\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:39 crc kubenswrapper[4787]: I1203 17:34:39.935353 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.038925 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.147841 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bl59w\" (UniqueName: \"kubernetes.io/projected/bd624460-116c-4630-b418-788f93ede485-kube-api-access-bl59w\") pod \"bd624460-116c-4630-b418-788f93ede485\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.148125 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-config\") pod \"bd624460-116c-4630-b418-788f93ede485\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.148155 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-dns-svc\") pod \"bd624460-116c-4630-b418-788f93ede485\" (UID: \"bd624460-116c-4630-b418-788f93ede485\") " Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.149000 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bd624460-116c-4630-b418-788f93ede485" (UID: "bd624460-116c-4630-b418-788f93ede485"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.152798 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-config" (OuterVolumeSpecName: "config") pod "bd624460-116c-4630-b418-788f93ede485" (UID: "bd624460-116c-4630-b418-788f93ede485"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.181499 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd624460-116c-4630-b418-788f93ede485-kube-api-access-bl59w" (OuterVolumeSpecName: "kube-api-access-bl59w") pod "bd624460-116c-4630-b418-788f93ede485" (UID: "bd624460-116c-4630-b418-788f93ede485"). InnerVolumeSpecName "kube-api-access-bl59w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.251062 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.251099 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bd624460-116c-4630-b418-788f93ede485-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.251111 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bl59w\" (UniqueName: \"kubernetes.io/projected/bd624460-116c-4630-b418-788f93ede485-kube-api-access-bl59w\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.428315 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-99x2s"] Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.610094 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.615481 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.618389 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.618581 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-fj8hk" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.618703 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.618909 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.622894 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" event={"ID":"f8dd3942-d40c-4699-8603-74f41dd659be","Type":"ContainerStarted","Data":"59e9a532f52466eaa748a292d5750bcea17561ee259a4354d8c98581845ee130"} Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.629682 4787 generic.go:334] "Generic (PLEG): container finished" podID="cab2a5b3-1157-45ab-ad4b-828a1638fd78" containerID="9fdfb1e0585fb00676e76659165c5315e8cf47e0535498c0f6c2ba0a7b67a3fe" exitCode=0 Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.629809 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mc6r6" event={"ID":"cab2a5b3-1157-45ab-ad4b-828a1638fd78","Type":"ContainerDied","Data":"9fdfb1e0585fb00676e76659165c5315e8cf47e0535498c0f6c2ba0a7b67a3fe"} Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.631788 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" event={"ID":"bd624460-116c-4630-b418-788f93ede485","Type":"ContainerDied","Data":"8c9064e9250e611b615be754744ef9580e57225dac0c1ca338c3e28cbaece8d6"} Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.631875 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-bmkgx" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.679989 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.758075 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkgx"] Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.762217 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/106d2d96-ed76-4c59-9577-7086e56f9878-cache\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.762346 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/106d2d96-ed76-4c59-9577-7086e56f9878-lock\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.762380 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.762465 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d6h99\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-kube-api-access-d6h99\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.762515 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.765512 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-bmkgx"] Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.799187 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.800135 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.815506 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.864109 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/106d2d96-ed76-4c59-9577-7086e56f9878-lock\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.864163 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.864243 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d6h99\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-kube-api-access-d6h99\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.864272 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.864397 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/106d2d96-ed76-4c59-9577-7086e56f9878-cache\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.864909 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.865871 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/106d2d96-ed76-4c59-9577-7086e56f9878-cache\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: E1203 17:34:40.866145 4787 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:34:40 crc kubenswrapper[4787]: E1203 17:34:40.866164 4787 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:34:40 crc kubenswrapper[4787]: E1203 17:34:40.866204 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift podName:106d2d96-ed76-4c59-9577-7086e56f9878 nodeName:}" failed. No retries permitted until 2025-12-03 17:34:41.366188519 +0000 UTC m=+1298.183659468 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift") pod "swift-storage-0" (UID: "106d2d96-ed76-4c59-9577-7086e56f9878") : configmap "swift-ring-files" not found Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.866297 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/106d2d96-ed76-4c59-9577-7086e56f9878-lock\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:40 crc kubenswrapper[4787]: I1203 17:34:40.975182 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d6h99\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-kube-api-access-d6h99\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:41 crc kubenswrapper[4787]: I1203 17:34:41.072923 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:41 crc kubenswrapper[4787]: I1203 17:34:41.412849 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:41 crc kubenswrapper[4787]: E1203 17:34:41.413618 4787 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:34:41 crc kubenswrapper[4787]: E1203 17:34:41.413648 4787 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:34:41 crc kubenswrapper[4787]: E1203 17:34:41.413700 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift podName:106d2d96-ed76-4c59-9577-7086e56f9878 nodeName:}" failed. No retries permitted until 2025-12-03 17:34:42.413682034 +0000 UTC m=+1299.231152983 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift") pod "swift-storage-0" (UID: "106d2d96-ed76-4c59-9577-7086e56f9878") : configmap "swift-ring-files" not found Dec 03 17:34:41 crc kubenswrapper[4787]: I1203 17:34:41.644620 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3b68d021-5db8-4a87-92f2-e827bbdd2255","Type":"ContainerStarted","Data":"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642"} Dec 03 17:34:41 crc kubenswrapper[4787]: I1203 17:34:41.648056 4787 generic.go:334] "Generic (PLEG): container finished" podID="f8dd3942-d40c-4699-8603-74f41dd659be" containerID="b9430b9948e102d90cbf4f27e499567757bd6ee342265a29b71aa0cd786b2512" exitCode=0 Dec 03 17:34:41 crc kubenswrapper[4787]: I1203 17:34:41.648179 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" event={"ID":"f8dd3942-d40c-4699-8603-74f41dd659be","Type":"ContainerDied","Data":"b9430b9948e102d90cbf4f27e499567757bd6ee342265a29b71aa0cd786b2512"} Dec 03 17:34:41 crc kubenswrapper[4787]: I1203 17:34:41.652551 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mc6r6" event={"ID":"cab2a5b3-1157-45ab-ad4b-828a1638fd78","Type":"ContainerStarted","Data":"3aecae32625db024ab01e744d7c60e52efba4e0200c9ed7b34cd88412b2193ba"} Dec 03 17:34:41 crc kubenswrapper[4787]: I1203 17:34:41.656636 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-6f77dc657b-v8bks" Dec 03 17:34:41 crc kubenswrapper[4787]: I1203 17:34:41.813328 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd624460-116c-4630-b418-788f93ede485" path="/var/lib/kubelet/pods/bd624460-116c-4630-b418-788f93ede485/volumes" Dec 03 17:34:41 crc kubenswrapper[4787]: I1203 17:34:41.814815 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-85474b5df5-nrlj4"] Dec 03 17:34:42 crc kubenswrapper[4787]: I1203 17:34:42.440308 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:42 crc kubenswrapper[4787]: E1203 17:34:42.440454 4787 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:34:42 crc kubenswrapper[4787]: E1203 17:34:42.440800 4787 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:34:42 crc kubenswrapper[4787]: E1203 17:34:42.440851 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift podName:106d2d96-ed76-4c59-9577-7086e56f9878 nodeName:}" failed. No retries permitted until 2025-12-03 17:34:44.440835064 +0000 UTC m=+1301.258306023 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift") pod "swift-storage-0" (UID: "106d2d96-ed76-4c59-9577-7086e56f9878") : configmap "swift-ring-files" not found Dec 03 17:34:42 crc kubenswrapper[4787]: I1203 17:34:42.664408 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" event={"ID":"f8dd3942-d40c-4699-8603-74f41dd659be","Type":"ContainerStarted","Data":"f64841e57f62de7bec4777a4961ba309046ab197325b6ec541c425aa885e28bb"} Dec 03 17:34:42 crc kubenswrapper[4787]: I1203 17:34:42.673150 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:42 crc kubenswrapper[4787]: I1203 17:34:42.678515 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-mc6r6" event={"ID":"cab2a5b3-1157-45ab-ad4b-828a1638fd78","Type":"ContainerStarted","Data":"6ba15209f9c69073c4fbda4090134b5beeee5e9fcfd51e08b3e1c77b2640b74a"} Dec 03 17:34:42 crc kubenswrapper[4787]: I1203 17:34:42.705138 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" podStartSLOduration=3.053605038 podStartE2EDuration="3.705071824s" podCreationTimestamp="2025-12-03 17:34:39 +0000 UTC" firstStartedPulling="2025-12-03 17:34:40.444419825 +0000 UTC m=+1297.261890784" lastFinishedPulling="2025-12-03 17:34:41.095886621 +0000 UTC m=+1297.913357570" observedRunningTime="2025-12-03 17:34:42.692557462 +0000 UTC m=+1299.510028421" watchObservedRunningTime="2025-12-03 17:34:42.705071824 +0000 UTC m=+1299.522542783" Dec 03 17:34:42 crc kubenswrapper[4787]: I1203 17:34:42.721409 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-mc6r6" podStartSLOduration=26.183099179 podStartE2EDuration="30.721388528s" podCreationTimestamp="2025-12-03 17:34:12 +0000 UTC" firstStartedPulling="2025-12-03 17:34:33.966598598 +0000 UTC m=+1290.784069557" lastFinishedPulling="2025-12-03 17:34:38.504887947 +0000 UTC m=+1295.322358906" observedRunningTime="2025-12-03 17:34:42.71507035 +0000 UTC m=+1299.532541319" watchObservedRunningTime="2025-12-03 17:34:42.721388528 +0000 UTC m=+1299.538859487" Dec 03 17:34:43 crc kubenswrapper[4787]: I1203 17:34:43.692491 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3180168d-9dad-441d-86bd-9d1801ef629f","Type":"ContainerStarted","Data":"2cad526e0647b21c366d8989df027fd6a0351ec3a648a935c487d2d379ae49af"} Dec 03 17:34:43 crc kubenswrapper[4787]: I1203 17:34:43.701303 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d73396bd-2e1f-4c20-8327-610c5d783373","Type":"ContainerStarted","Data":"e140a21aafabe034cbd6828830bc11cbed6f27b23fdfcc9449ef640e44a0560b"} Dec 03 17:34:43 crc kubenswrapper[4787]: I1203 17:34:43.702504 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:43 crc kubenswrapper[4787]: I1203 17:34:43.702560 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.496513 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:44 crc kubenswrapper[4787]: E1203 17:34:44.496681 4787 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:34:44 crc kubenswrapper[4787]: E1203 17:34:44.496922 4787 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:34:44 crc kubenswrapper[4787]: E1203 17:34:44.496985 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift podName:106d2d96-ed76-4c59-9577-7086e56f9878 nodeName:}" failed. No retries permitted until 2025-12-03 17:34:48.496965842 +0000 UTC m=+1305.314436801 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift") pod "swift-storage-0" (UID: "106d2d96-ed76-4c59-9577-7086e56f9878") : configmap "swift-ring-files" not found Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.651862 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-4fpwj"] Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.653954 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.662618 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.662957 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.663217 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.675564 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-4fpwj"] Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.802987 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-dispersionconf\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.803164 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-scripts\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.803230 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-ring-data-devices\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.803585 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n2wts\" (UniqueName: \"kubernetes.io/projected/973215cc-7718-421c-b374-49e64bb9af3e-kube-api-access-n2wts\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.803708 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-combined-ca-bundle\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.803903 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-swiftconf\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.804194 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/973215cc-7718-421c-b374-49e64bb9af3e-etc-swift\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.906369 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-scripts\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.906919 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-ring-data-devices\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.907996 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n2wts\" (UniqueName: \"kubernetes.io/projected/973215cc-7718-421c-b374-49e64bb9af3e-kube-api-access-n2wts\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.909409 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-ring-data-devices\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.909430 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-scripts\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.910984 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-combined-ca-bundle\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.912210 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-swiftconf\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.912388 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/973215cc-7718-421c-b374-49e64bb9af3e-etc-swift\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.912504 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-dispersionconf\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.913388 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/973215cc-7718-421c-b374-49e64bb9af3e-etc-swift\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.919709 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-combined-ca-bundle\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.924730 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-dispersionconf\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.925412 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-swiftconf\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.930202 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n2wts\" (UniqueName: \"kubernetes.io/projected/973215cc-7718-421c-b374-49e64bb9af3e-kube-api-access-n2wts\") pod \"swift-ring-rebalance-4fpwj\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:44 crc kubenswrapper[4787]: I1203 17:34:44.994480 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:34:45 crc kubenswrapper[4787]: I1203 17:34:45.486099 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-4fpwj"] Dec 03 17:34:45 crc kubenswrapper[4787]: I1203 17:34:45.746488 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4fpwj" event={"ID":"973215cc-7718-421c-b374-49e64bb9af3e","Type":"ContainerStarted","Data":"0a469acb7fca797ef599a0e1917fb95f46bef8eb8c15499bcb29111d2e56e19a"} Dec 03 17:34:45 crc kubenswrapper[4787]: I1203 17:34:45.748950 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"6b71b702-ef57-41ac-8ada-1fe782d5092a","Type":"ContainerStarted","Data":"3e4acb569b024047d1553d11dc61367d2c8e4ef65e7235a63f8888ef49c7e457"} Dec 03 17:34:45 crc kubenswrapper[4787]: I1203 17:34:45.752060 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"cd30e7ed-2464-4857-a563-b02f14871d54","Type":"ContainerStarted","Data":"b2a9d9a7b60a94b927f9322f5efa6286671de2fe790195558f0a53e55363e126"} Dec 03 17:34:45 crc kubenswrapper[4787]: I1203 17:34:45.777861 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=19.778083149 podStartE2EDuration="31.777838092s" podCreationTimestamp="2025-12-03 17:34:14 +0000 UTC" firstStartedPulling="2025-12-03 17:34:32.80742642 +0000 UTC m=+1289.624897379" lastFinishedPulling="2025-12-03 17:34:44.807181343 +0000 UTC m=+1301.624652322" observedRunningTime="2025-12-03 17:34:45.76948551 +0000 UTC m=+1302.586956469" watchObservedRunningTime="2025-12-03 17:34:45.777838092 +0000 UTC m=+1302.595309071" Dec 03 17:34:45 crc kubenswrapper[4787]: I1203 17:34:45.783831 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:45 crc kubenswrapper[4787]: I1203 17:34:45.783886 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:45 crc kubenswrapper[4787]: I1203 17:34:45.796589 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=21.442306575 podStartE2EDuration="31.79656671s" podCreationTimestamp="2025-12-03 17:34:14 +0000 UTC" firstStartedPulling="2025-12-03 17:34:34.470726901 +0000 UTC m=+1291.288197860" lastFinishedPulling="2025-12-03 17:34:44.824987036 +0000 UTC m=+1301.642457995" observedRunningTime="2025-12-03 17:34:45.792833931 +0000 UTC m=+1302.610304900" watchObservedRunningTime="2025-12-03 17:34:45.79656671 +0000 UTC m=+1302.614037669" Dec 03 17:34:45 crc kubenswrapper[4787]: I1203 17:34:45.848357 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:46 crc kubenswrapper[4787]: I1203 17:34:46.151359 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:46 crc kubenswrapper[4787]: I1203 17:34:46.151871 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:46 crc kubenswrapper[4787]: I1203 17:34:46.206433 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:46 crc kubenswrapper[4787]: I1203 17:34:46.298855 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 03 17:34:46 crc kubenswrapper[4787]: I1203 17:34:46.298912 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 03 17:34:46 crc kubenswrapper[4787]: I1203 17:34:46.376581 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 03 17:34:46 crc kubenswrapper[4787]: I1203 17:34:46.873052 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 03 17:34:46 crc kubenswrapper[4787]: I1203 17:34:46.896444 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 03 17:34:46 crc kubenswrapper[4787]: I1203 17:34:46.945688 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.132376 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-dxzmh"] Dec 03 17:34:47 crc kubenswrapper[4787]: E1203 17:34:47.134213 4787 log.go:32] "CreateContainer in sandbox from runtime service failed" err=< Dec 03 17:34:47 crc kubenswrapper[4787]: rpc error: code = Unknown desc = container create failed: mount `/var/lib/kubelet/pods/5dc53172-a8d8-44f8-9969-d5ed647dcd71/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 17:34:47 crc kubenswrapper[4787]: > podSandboxID="1cc0d2397e8835be72f1670cc38afcb4010d60869a22f58b59b7d9aef0ec825f" Dec 03 17:34:47 crc kubenswrapper[4787]: E1203 17:34:47.134356 4787 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 03 17:34:47 crc kubenswrapper[4787]: init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nfdh5dfhb6h64h676hc4h78h97h669h54chfbh696hb5h54bh5d4h6bh64h644h677h584h5cbh698h9dh5bbh5f8h5b8hcdh644h5c7h694hbfh589q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nw9ms,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5ccc8479f9-dxzmh_openstack(5dc53172-a8d8-44f8-9969-d5ed647dcd71): CreateContainerError: container create failed: mount `/var/lib/kubelet/pods/5dc53172-a8d8-44f8-9969-d5ed647dcd71/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory Dec 03 17:34:47 crc kubenswrapper[4787]: > logger="UnhandledError" Dec 03 17:34:47 crc kubenswrapper[4787]: E1203 17:34:47.138783 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with CreateContainerError: \"container create failed: mount `/var/lib/kubelet/pods/5dc53172-a8d8-44f8-9969-d5ed647dcd71/volume-subpaths/dns-svc/init/1` to `etc/dnsmasq.d/hosts/dns-svc`: No such file or directory\\n\"" pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" podUID="5dc53172-a8d8-44f8-9969-d5ed647dcd71" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.179002 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-7kjsl"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.190576 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.204254 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.242910 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-7kjsl"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.279106 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.279539 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnrbz\" (UniqueName: \"kubernetes.io/projected/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-kube-api-access-qnrbz\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.279639 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-config\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.279740 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.292132 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-mbdkp"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.293675 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.297351 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.314676 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-mbdkp"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.388055 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cbb0a5a5-509c-4786-8225-22ed73d9ab20-ovn-rundir\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.388147 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cbb0a5a5-509c-4786-8225-22ed73d9ab20-ovs-rundir\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.388196 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbb0a5a5-509c-4786-8225-22ed73d9ab20-config\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.388255 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.388459 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbb0a5a5-509c-4786-8225-22ed73d9ab20-combined-ca-bundle\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.388589 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbb0a5a5-509c-4786-8225-22ed73d9ab20-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.388791 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnrbz\" (UniqueName: \"kubernetes.io/projected/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-kube-api-access-qnrbz\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.388891 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-config\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.388949 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7gxv\" (UniqueName: \"kubernetes.io/projected/cbb0a5a5-509c-4786-8225-22ed73d9ab20-kube-api-access-m7gxv\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.389045 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.390811 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-config\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.391453 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-ovsdbserver-sb\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.392438 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-dns-svc\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.416219 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8ee8-account-create-update-bvmfg"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.437805 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8ee8-account-create-update-bvmfg" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.445185 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.467068 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8ee8-account-create-update-bvmfg"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.483712 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnrbz\" (UniqueName: \"kubernetes.io/projected/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-kube-api-access-qnrbz\") pod \"dnsmasq-dns-6c89d5d749-7kjsl\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.505052 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5spth\" (UniqueName: \"kubernetes.io/projected/2131a38e-a365-46fd-8b7f-b01906edec67-kube-api-access-5spth\") pod \"keystone-8ee8-account-create-update-bvmfg\" (UID: \"2131a38e-a365-46fd-8b7f-b01906edec67\") " pod="openstack/keystone-8ee8-account-create-update-bvmfg" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.505457 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7gxv\" (UniqueName: \"kubernetes.io/projected/cbb0a5a5-509c-4786-8225-22ed73d9ab20-kube-api-access-m7gxv\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.505498 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2131a38e-a365-46fd-8b7f-b01906edec67-operator-scripts\") pod \"keystone-8ee8-account-create-update-bvmfg\" (UID: \"2131a38e-a365-46fd-8b7f-b01906edec67\") " pod="openstack/keystone-8ee8-account-create-update-bvmfg" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.505603 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cbb0a5a5-509c-4786-8225-22ed73d9ab20-ovn-rundir\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.505640 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cbb0a5a5-509c-4786-8225-22ed73d9ab20-ovs-rundir\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.505718 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbb0a5a5-509c-4786-8225-22ed73d9ab20-config\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.505816 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbb0a5a5-509c-4786-8225-22ed73d9ab20-combined-ca-bundle\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.505864 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbb0a5a5-509c-4786-8225-22ed73d9ab20-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.507009 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/cbb0a5a5-509c-4786-8225-22ed73d9ab20-ovn-rundir\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.509934 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbb0a5a5-509c-4786-8225-22ed73d9ab20-config\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.510030 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/cbb0a5a5-509c-4786-8225-22ed73d9ab20-ovs-rundir\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.512442 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-h5xrc"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.516195 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cbb0a5a5-509c-4786-8225-22ed73d9ab20-combined-ca-bundle\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.518555 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-h5xrc" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.533767 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.535271 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/cbb0a5a5-509c-4786-8225-22ed73d9ab20-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.550133 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.552500 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.560573 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7gxv\" (UniqueName: \"kubernetes.io/projected/cbb0a5a5-509c-4786-8225-22ed73d9ab20-kube-api-access-m7gxv\") pod \"ovn-controller-metrics-mbdkp\" (UID: \"cbb0a5a5-509c-4786-8225-22ed73d9ab20\") " pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.560884 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.561181 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.561343 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-f4zwv" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.561483 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.584409 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-h5xrc"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.606141 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-99x2s"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.606535 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" podUID="f8dd3942-d40c-4699-8603-74f41dd659be" containerName="dnsmasq-dns" containerID="cri-o://f64841e57f62de7bec4777a4961ba309046ab197325b6ec541c425aa885e28bb" gracePeriod=10 Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.608817 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.624650 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-mbdkp" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.625904 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5spth\" (UniqueName: \"kubernetes.io/projected/2131a38e-a365-46fd-8b7f-b01906edec67-kube-api-access-5spth\") pod \"keystone-8ee8-account-create-update-bvmfg\" (UID: \"2131a38e-a365-46fd-8b7f-b01906edec67\") " pod="openstack/keystone-8ee8-account-create-update-bvmfg" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.626053 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2131a38e-a365-46fd-8b7f-b01906edec67-operator-scripts\") pod \"keystone-8ee8-account-create-update-bvmfg\" (UID: \"2131a38e-a365-46fd-8b7f-b01906edec67\") " pod="openstack/keystone-8ee8-account-create-update-bvmfg" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.626209 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bccb94-fcb5-44fc-819b-e39562a7e57a-operator-scripts\") pod \"placement-db-create-h5xrc\" (UID: \"d3bccb94-fcb5-44fc-819b-e39562a7e57a\") " pod="openstack/placement-db-create-h5xrc" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.626326 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxszf\" (UniqueName: \"kubernetes.io/projected/d3bccb94-fcb5-44fc-819b-e39562a7e57a-kube-api-access-jxszf\") pod \"placement-db-create-h5xrc\" (UID: \"d3bccb94-fcb5-44fc-819b-e39562a7e57a\") " pod="openstack/placement-db-create-h5xrc" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.627430 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2131a38e-a365-46fd-8b7f-b01906edec67-operator-scripts\") pod \"keystone-8ee8-account-create-update-bvmfg\" (UID: \"2131a38e-a365-46fd-8b7f-b01906edec67\") " pod="openstack/keystone-8ee8-account-create-update-bvmfg" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.639895 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.662443 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5spth\" (UniqueName: \"kubernetes.io/projected/2131a38e-a365-46fd-8b7f-b01906edec67-kube-api-access-5spth\") pod \"keystone-8ee8-account-create-update-bvmfg\" (UID: \"2131a38e-a365-46fd-8b7f-b01906edec67\") " pod="openstack/keystone-8ee8-account-create-update-bvmfg" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.666293 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-698758b865-kspp6"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.670972 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.676334 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.693660 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.693709 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.722376 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kspp6"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.732618 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/df89848b-864f-477f-bea0-e34cfcaba0fb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.732760 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df89848b-864f-477f-bea0-e34cfcaba0fb-scripts\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.732805 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df89848b-864f-477f-bea0-e34cfcaba0fb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.732835 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcffk\" (UniqueName: \"kubernetes.io/projected/df89848b-864f-477f-bea0-e34cfcaba0fb-kube-api-access-rcffk\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.732870 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/df89848b-864f-477f-bea0-e34cfcaba0fb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.732922 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df89848b-864f-477f-bea0-e34cfcaba0fb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.732973 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bccb94-fcb5-44fc-819b-e39562a7e57a-operator-scripts\") pod \"placement-db-create-h5xrc\" (UID: \"d3bccb94-fcb5-44fc-819b-e39562a7e57a\") " pod="openstack/placement-db-create-h5xrc" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.733060 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxszf\" (UniqueName: \"kubernetes.io/projected/d3bccb94-fcb5-44fc-819b-e39562a7e57a-kube-api-access-jxszf\") pod \"placement-db-create-h5xrc\" (UID: \"d3bccb94-fcb5-44fc-819b-e39562a7e57a\") " pod="openstack/placement-db-create-h5xrc" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.733092 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df89848b-864f-477f-bea0-e34cfcaba0fb-config\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.737860 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bccb94-fcb5-44fc-819b-e39562a7e57a-operator-scripts\") pod \"placement-db-create-h5xrc\" (UID: \"d3bccb94-fcb5-44fc-819b-e39562a7e57a\") " pod="openstack/placement-db-create-h5xrc" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.743570 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-a415-account-create-update-t8r4c"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.767274 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxszf\" (UniqueName: \"kubernetes.io/projected/d3bccb94-fcb5-44fc-819b-e39562a7e57a-kube-api-access-jxszf\") pod \"placement-db-create-h5xrc\" (UID: \"d3bccb94-fcb5-44fc-819b-e39562a7e57a\") " pod="openstack/placement-db-create-h5xrc" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.770183 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a415-account-create-update-t8r4c" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.775125 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.824912 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a415-account-create-update-t8r4c"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.835308 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.835668 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/df89848b-864f-477f-bea0-e34cfcaba0fb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.835788 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfrnl\" (UniqueName: \"kubernetes.io/projected/6c6d0805-aa7a-4038-8245-77e79adbe026-kube-api-access-cfrnl\") pod \"placement-a415-account-create-update-t8r4c\" (UID: \"6c6d0805-aa7a-4038-8245-77e79adbe026\") " pod="openstack/placement-a415-account-create-update-t8r4c" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.835884 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-config\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.835963 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgdkf\" (UniqueName: \"kubernetes.io/projected/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-kube-api-access-bgdkf\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.836083 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df89848b-864f-477f-bea0-e34cfcaba0fb-scripts\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.836182 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-dns-svc\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.836265 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df89848b-864f-477f-bea0-e34cfcaba0fb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.836350 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcffk\" (UniqueName: \"kubernetes.io/projected/df89848b-864f-477f-bea0-e34cfcaba0fb-kube-api-access-rcffk\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.836477 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c6d0805-aa7a-4038-8245-77e79adbe026-operator-scripts\") pod \"placement-a415-account-create-update-t8r4c\" (UID: \"6c6d0805-aa7a-4038-8245-77e79adbe026\") " pod="openstack/placement-a415-account-create-update-t8r4c" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.836626 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/df89848b-864f-477f-bea0-e34cfcaba0fb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.836786 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df89848b-864f-477f-bea0-e34cfcaba0fb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.836928 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.837216 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df89848b-864f-477f-bea0-e34cfcaba0fb-config\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.837722 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/df89848b-864f-477f-bea0-e34cfcaba0fb-scripts\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.842207 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/df89848b-864f-477f-bea0-e34cfcaba0fb-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.844430 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/df89848b-864f-477f-bea0-e34cfcaba0fb-config\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.850937 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/df89848b-864f-477f-bea0-e34cfcaba0fb-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.870852 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/df89848b-864f-477f-bea0-e34cfcaba0fb-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.876253 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8ee8-account-create-update-bvmfg" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.879433 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.879770 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcffk\" (UniqueName: \"kubernetes.io/projected/df89848b-864f-477f-bea0-e34cfcaba0fb-kube-api-access-rcffk\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.882144 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df89848b-864f-477f-bea0-e34cfcaba0fb-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"df89848b-864f-477f-bea0-e34cfcaba0fb\") " pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.898982 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-h5xrc" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.913365 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-z6tr4"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.918512 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-z6tr4" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.939198 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-z6tr4"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.940040 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.941034 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-dns-svc\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.941118 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c6d0805-aa7a-4038-8245-77e79adbe026-operator-scripts\") pod \"placement-a415-account-create-update-t8r4c\" (UID: \"6c6d0805-aa7a-4038-8245-77e79adbe026\") " pod="openstack/placement-a415-account-create-update-t8r4c" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.941183 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.941253 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.941322 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfrnl\" (UniqueName: \"kubernetes.io/projected/6c6d0805-aa7a-4038-8245-77e79adbe026-kube-api-access-cfrnl\") pod \"placement-a415-account-create-update-t8r4c\" (UID: \"6c6d0805-aa7a-4038-8245-77e79adbe026\") " pod="openstack/placement-a415-account-create-update-t8r4c" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.941345 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-config\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.941366 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgdkf\" (UniqueName: \"kubernetes.io/projected/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-kube-api-access-bgdkf\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.944046 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c6d0805-aa7a-4038-8245-77e79adbe026-operator-scripts\") pod \"placement-a415-account-create-update-t8r4c\" (UID: \"6c6d0805-aa7a-4038-8245-77e79adbe026\") " pod="openstack/placement-a415-account-create-update-t8r4c" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.945108 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-sb\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.945593 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-config\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.945701 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-nb\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.958096 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-dns-svc\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.969584 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfrnl\" (UniqueName: \"kubernetes.io/projected/6c6d0805-aa7a-4038-8245-77e79adbe026-kube-api-access-cfrnl\") pod \"placement-a415-account-create-update-t8r4c\" (UID: \"6c6d0805-aa7a-4038-8245-77e79adbe026\") " pod="openstack/placement-a415-account-create-update-t8r4c" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.979313 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-6bd0-account-create-update-qvtnl"] Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.981307 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6bd0-account-create-update-qvtnl" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.981596 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgdkf\" (UniqueName: \"kubernetes.io/projected/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-kube-api-access-bgdkf\") pod \"dnsmasq-dns-698758b865-kspp6\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:47 crc kubenswrapper[4787]: I1203 17:34:47.991911 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:47.999819 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6bd0-account-create-update-qvtnl"] Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.045709 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.048388 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40edf4b6-dc92-4d34-bbb2-a3784cf06405-operator-scripts\") pod \"glance-db-create-z6tr4\" (UID: \"40edf4b6-dc92-4d34-bbb2-a3784cf06405\") " pod="openstack/glance-db-create-z6tr4" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.049298 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pr6mx\" (UniqueName: \"kubernetes.io/projected/40edf4b6-dc92-4d34-bbb2-a3784cf06405-kube-api-access-pr6mx\") pod \"glance-db-create-z6tr4\" (UID: \"40edf4b6-dc92-4d34-bbb2-a3784cf06405\") " pod="openstack/glance-db-create-z6tr4" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.101711 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.110643 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a415-account-create-update-t8r4c" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.162285 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad8fa82a-ce58-4713-900e-f0e333cd8448-operator-scripts\") pod \"glance-6bd0-account-create-update-qvtnl\" (UID: \"ad8fa82a-ce58-4713-900e-f0e333cd8448\") " pod="openstack/glance-6bd0-account-create-update-qvtnl" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.162382 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40edf4b6-dc92-4d34-bbb2-a3784cf06405-operator-scripts\") pod \"glance-db-create-z6tr4\" (UID: \"40edf4b6-dc92-4d34-bbb2-a3784cf06405\") " pod="openstack/glance-db-create-z6tr4" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.162421 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd5kx\" (UniqueName: \"kubernetes.io/projected/ad8fa82a-ce58-4713-900e-f0e333cd8448-kube-api-access-gd5kx\") pod \"glance-6bd0-account-create-update-qvtnl\" (UID: \"ad8fa82a-ce58-4713-900e-f0e333cd8448\") " pod="openstack/glance-6bd0-account-create-update-qvtnl" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.162591 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pr6mx\" (UniqueName: \"kubernetes.io/projected/40edf4b6-dc92-4d34-bbb2-a3784cf06405-kube-api-access-pr6mx\") pod \"glance-db-create-z6tr4\" (UID: \"40edf4b6-dc92-4d34-bbb2-a3784cf06405\") " pod="openstack/glance-db-create-z6tr4" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.164865 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40edf4b6-dc92-4d34-bbb2-a3784cf06405-operator-scripts\") pod \"glance-db-create-z6tr4\" (UID: \"40edf4b6-dc92-4d34-bbb2-a3784cf06405\") " pod="openstack/glance-db-create-z6tr4" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.238067 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pr6mx\" (UniqueName: \"kubernetes.io/projected/40edf4b6-dc92-4d34-bbb2-a3784cf06405-kube-api-access-pr6mx\") pod \"glance-db-create-z6tr4\" (UID: \"40edf4b6-dc92-4d34-bbb2-a3784cf06405\") " pod="openstack/glance-db-create-z6tr4" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.291927 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad8fa82a-ce58-4713-900e-f0e333cd8448-operator-scripts\") pod \"glance-6bd0-account-create-update-qvtnl\" (UID: \"ad8fa82a-ce58-4713-900e-f0e333cd8448\") " pod="openstack/glance-6bd0-account-create-update-qvtnl" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.292568 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd5kx\" (UniqueName: \"kubernetes.io/projected/ad8fa82a-ce58-4713-900e-f0e333cd8448-kube-api-access-gd5kx\") pod \"glance-6bd0-account-create-update-qvtnl\" (UID: \"ad8fa82a-ce58-4713-900e-f0e333cd8448\") " pod="openstack/glance-6bd0-account-create-update-qvtnl" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.313159 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad8fa82a-ce58-4713-900e-f0e333cd8448-operator-scripts\") pod \"glance-6bd0-account-create-update-qvtnl\" (UID: \"ad8fa82a-ce58-4713-900e-f0e333cd8448\") " pod="openstack/glance-6bd0-account-create-update-qvtnl" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.346922 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd5kx\" (UniqueName: \"kubernetes.io/projected/ad8fa82a-ce58-4713-900e-f0e333cd8448-kube-api-access-gd5kx\") pod \"glance-6bd0-account-create-update-qvtnl\" (UID: \"ad8fa82a-ce58-4713-900e-f0e333cd8448\") " pod="openstack/glance-6bd0-account-create-update-qvtnl" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.460498 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-z6tr4" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.499852 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:48 crc kubenswrapper[4787]: E1203 17:34:48.500417 4787 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:34:48 crc kubenswrapper[4787]: E1203 17:34:48.500441 4787 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:34:48 crc kubenswrapper[4787]: E1203 17:34:48.500511 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift podName:106d2d96-ed76-4c59-9577-7086e56f9878 nodeName:}" failed. No retries permitted until 2025-12-03 17:34:56.500488748 +0000 UTC m=+1313.317959707 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift") pod "swift-storage-0" (UID: "106d2d96-ed76-4c59-9577-7086e56f9878") : configmap "swift-ring-files" not found Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.515560 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6bd0-account-create-update-qvtnl" Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.831880 4787 generic.go:334] "Generic (PLEG): container finished" podID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerID="eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642" exitCode=0 Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.831964 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3b68d021-5db8-4a87-92f2-e827bbdd2255","Type":"ContainerDied","Data":"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642"} Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.846585 4787 generic.go:334] "Generic (PLEG): container finished" podID="f8dd3942-d40c-4699-8603-74f41dd659be" containerID="f64841e57f62de7bec4777a4961ba309046ab197325b6ec541c425aa885e28bb" exitCode=0 Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.846705 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" event={"ID":"f8dd3942-d40c-4699-8603-74f41dd659be","Type":"ContainerDied","Data":"f64841e57f62de7bec4777a4961ba309046ab197325b6ec541c425aa885e28bb"} Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.893320 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-mbdkp"] Dec 03 17:34:48 crc kubenswrapper[4787]: I1203 17:34:48.993937 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-7kjsl"] Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.258592 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-xh2lw"] Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.261031 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.266386 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-xh2lw"] Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.336115 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb521dcb-0316-4363-983f-1e2800ee2630-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-xh2lw\" (UID: \"fb521dcb-0316-4363-983f-1e2800ee2630\") " pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.336170 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gh54\" (UniqueName: \"kubernetes.io/projected/fb521dcb-0316-4363-983f-1e2800ee2630-kube-api-access-6gh54\") pod \"mysqld-exporter-openstack-db-create-xh2lw\" (UID: \"fb521dcb-0316-4363-983f-1e2800ee2630\") " pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.439231 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb521dcb-0316-4363-983f-1e2800ee2630-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-xh2lw\" (UID: \"fb521dcb-0316-4363-983f-1e2800ee2630\") " pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.439286 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gh54\" (UniqueName: \"kubernetes.io/projected/fb521dcb-0316-4363-983f-1e2800ee2630-kube-api-access-6gh54\") pod \"mysqld-exporter-openstack-db-create-xh2lw\" (UID: \"fb521dcb-0316-4363-983f-1e2800ee2630\") " pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.440565 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb521dcb-0316-4363-983f-1e2800ee2630-operator-scripts\") pod \"mysqld-exporter-openstack-db-create-xh2lw\" (UID: \"fb521dcb-0316-4363-983f-1e2800ee2630\") " pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.471838 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gh54\" (UniqueName: \"kubernetes.io/projected/fb521dcb-0316-4363-983f-1e2800ee2630-kube-api-access-6gh54\") pod \"mysqld-exporter-openstack-db-create-xh2lw\" (UID: \"fb521dcb-0316-4363-983f-1e2800ee2630\") " pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.473533 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-acb9-account-create-update-wv7cq"] Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.475539 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.477774 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-db-secret" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.488703 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-acb9-account-create-update-wv7cq"] Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.505875 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.542347 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f041eda1-f785-472b-928e-7ab2b901e99e-operator-scripts\") pod \"mysqld-exporter-acb9-account-create-update-wv7cq\" (UID: \"f041eda1-f785-472b-928e-7ab2b901e99e\") " pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.542687 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdq78\" (UniqueName: \"kubernetes.io/projected/f041eda1-f785-472b-928e-7ab2b901e99e-kube-api-access-mdq78\") pod \"mysqld-exporter-acb9-account-create-update-wv7cq\" (UID: \"f041eda1-f785-472b-928e-7ab2b901e99e\") " pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.589698 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.648606 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f041eda1-f785-472b-928e-7ab2b901e99e-operator-scripts\") pod \"mysqld-exporter-acb9-account-create-update-wv7cq\" (UID: \"f041eda1-f785-472b-928e-7ab2b901e99e\") " pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.649292 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f041eda1-f785-472b-928e-7ab2b901e99e-operator-scripts\") pod \"mysqld-exporter-acb9-account-create-update-wv7cq\" (UID: \"f041eda1-f785-472b-928e-7ab2b901e99e\") " pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.649506 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdq78\" (UniqueName: \"kubernetes.io/projected/f041eda1-f785-472b-928e-7ab2b901e99e-kube-api-access-mdq78\") pod \"mysqld-exporter-acb9-account-create-update-wv7cq\" (UID: \"f041eda1-f785-472b-928e-7ab2b901e99e\") " pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.668002 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdq78\" (UniqueName: \"kubernetes.io/projected/f041eda1-f785-472b-928e-7ab2b901e99e-kube-api-access-mdq78\") pod \"mysqld-exporter-acb9-account-create-update-wv7cq\" (UID: \"f041eda1-f785-472b-928e-7ab2b901e99e\") " pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" Dec 03 17:34:49 crc kubenswrapper[4787]: I1203 17:34:49.839785 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.127557 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.133415 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.238624 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-config\") pod \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.238673 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9nzmp\" (UniqueName: \"kubernetes.io/projected/f8dd3942-d40c-4699-8603-74f41dd659be-kube-api-access-9nzmp\") pod \"f8dd3942-d40c-4699-8603-74f41dd659be\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.238779 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-config\") pod \"f8dd3942-d40c-4699-8603-74f41dd659be\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.238852 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nw9ms\" (UniqueName: \"kubernetes.io/projected/5dc53172-a8d8-44f8-9969-d5ed647dcd71-kube-api-access-nw9ms\") pod \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.238883 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-dns-svc\") pod \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\" (UID: \"5dc53172-a8d8-44f8-9969-d5ed647dcd71\") " Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.238951 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-dns-svc\") pod \"f8dd3942-d40c-4699-8603-74f41dd659be\" (UID: \"f8dd3942-d40c-4699-8603-74f41dd659be\") " Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.244259 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5dc53172-a8d8-44f8-9969-d5ed647dcd71-kube-api-access-nw9ms" (OuterVolumeSpecName: "kube-api-access-nw9ms") pod "5dc53172-a8d8-44f8-9969-d5ed647dcd71" (UID: "5dc53172-a8d8-44f8-9969-d5ed647dcd71"). InnerVolumeSpecName "kube-api-access-nw9ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.244803 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8dd3942-d40c-4699-8603-74f41dd659be-kube-api-access-9nzmp" (OuterVolumeSpecName: "kube-api-access-9nzmp") pod "f8dd3942-d40c-4699-8603-74f41dd659be" (UID: "f8dd3942-d40c-4699-8603-74f41dd659be"). InnerVolumeSpecName "kube-api-access-9nzmp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.280456 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5dc53172-a8d8-44f8-9969-d5ed647dcd71" (UID: "5dc53172-a8d8-44f8-9969-d5ed647dcd71"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.304551 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-config" (OuterVolumeSpecName: "config") pod "5dc53172-a8d8-44f8-9969-d5ed647dcd71" (UID: "5dc53172-a8d8-44f8-9969-d5ed647dcd71"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.315074 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f8dd3942-d40c-4699-8603-74f41dd659be" (UID: "f8dd3942-d40c-4699-8603-74f41dd659be"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.316187 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-config" (OuterVolumeSpecName: "config") pod "f8dd3942-d40c-4699-8603-74f41dd659be" (UID: "f8dd3942-d40c-4699-8603-74f41dd659be"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.344158 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nw9ms\" (UniqueName: \"kubernetes.io/projected/5dc53172-a8d8-44f8-9969-d5ed647dcd71-kube-api-access-nw9ms\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.344227 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.344238 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.344246 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5dc53172-a8d8-44f8-9969-d5ed647dcd71-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.344371 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9nzmp\" (UniqueName: \"kubernetes.io/projected/f8dd3942-d40c-4699-8603-74f41dd659be-kube-api-access-9nzmp\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.344557 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f8dd3942-d40c-4699-8603-74f41dd659be-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.919299 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" event={"ID":"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf","Type":"ContainerStarted","Data":"f4fa2a31d0a4bcdf092be33a1bf0cbd3126fe1e1fe9ff86c111f873f08c66efb"} Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.921088 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" event={"ID":"f8dd3942-d40c-4699-8603-74f41dd659be","Type":"ContainerDied","Data":"59e9a532f52466eaa748a292d5750bcea17561ee259a4354d8c98581845ee130"} Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.921146 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.921159 4787 scope.go:117] "RemoveContainer" containerID="f64841e57f62de7bec4777a4961ba309046ab197325b6ec541c425aa885e28bb" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.922775 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.922779 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-dxzmh" event={"ID":"5dc53172-a8d8-44f8-9969-d5ed647dcd71","Type":"ContainerDied","Data":"1cc0d2397e8835be72f1670cc38afcb4010d60869a22f58b59b7d9aef0ec825f"} Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.924336 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-mbdkp" event={"ID":"cbb0a5a5-509c-4786-8225-22ed73d9ab20","Type":"ContainerStarted","Data":"43c92fb0a7a0ba30a3793343225b43d4a38209fefd95057606960a4791d01235"} Dec 03 17:34:54 crc kubenswrapper[4787]: I1203 17:34:54.936963 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7cb5889db5-99x2s" podUID="f8dd3942-d40c-4699-8603-74f41dd659be" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.132:5353: i/o timeout" Dec 03 17:34:55 crc kubenswrapper[4787]: I1203 17:34:55.006073 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-dxzmh"] Dec 03 17:34:55 crc kubenswrapper[4787]: I1203 17:34:55.025629 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-dxzmh"] Dec 03 17:34:55 crc kubenswrapper[4787]: I1203 17:34:55.035529 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-99x2s"] Dec 03 17:34:55 crc kubenswrapper[4787]: I1203 17:34:55.042583 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7cb5889db5-99x2s"] Dec 03 17:34:55 crc kubenswrapper[4787]: I1203 17:34:55.780178 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5dc53172-a8d8-44f8-9969-d5ed647dcd71" path="/var/lib/kubelet/pods/5dc53172-a8d8-44f8-9969-d5ed647dcd71/volumes" Dec 03 17:34:55 crc kubenswrapper[4787]: I1203 17:34:55.781083 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8dd3942-d40c-4699-8603-74f41dd659be" path="/var/lib/kubelet/pods/f8dd3942-d40c-4699-8603-74f41dd659be/volumes" Dec 03 17:34:56 crc kubenswrapper[4787]: I1203 17:34:56.033163 4787 scope.go:117] "RemoveContainer" containerID="b9430b9948e102d90cbf4f27e499567757bd6ee342265a29b71aa0cd786b2512" Dec 03 17:34:56 crc kubenswrapper[4787]: I1203 17:34:56.585808 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:34:56 crc kubenswrapper[4787]: E1203 17:34:56.586711 4787 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Dec 03 17:34:56 crc kubenswrapper[4787]: E1203 17:34:56.586729 4787 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Dec 03 17:34:56 crc kubenswrapper[4787]: E1203 17:34:56.586779 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift podName:106d2d96-ed76-4c59-9577-7086e56f9878 nodeName:}" failed. No retries permitted until 2025-12-03 17:35:12.586762246 +0000 UTC m=+1329.404233205 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift") pod "swift-storage-0" (UID: "106d2d96-ed76-4c59-9577-7086e56f9878") : configmap "swift-ring-files" not found Dec 03 17:34:56 crc kubenswrapper[4787]: I1203 17:34:56.698996 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8ee8-account-create-update-bvmfg"] Dec 03 17:34:56 crc kubenswrapper[4787]: I1203 17:34:56.707612 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-6bd0-account-create-update-qvtnl"] Dec 03 17:34:56 crc kubenswrapper[4787]: I1203 17:34:56.716608 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-a415-account-create-update-t8r4c"] Dec 03 17:34:56 crc kubenswrapper[4787]: W1203 17:34:56.771147 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2131a38e_a365_46fd_8b7f_b01906edec67.slice/crio-85a879c3cdde78dee6e1f2a07b37049b87a34640339f0d4200c9e16aef4c9f9b WatchSource:0}: Error finding container 85a879c3cdde78dee6e1f2a07b37049b87a34640339f0d4200c9e16aef4c9f9b: Status 404 returned error can't find the container with id 85a879c3cdde78dee6e1f2a07b37049b87a34640339f0d4200c9e16aef4c9f9b Dec 03 17:34:56 crc kubenswrapper[4787]: W1203 17:34:56.775604 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6c6d0805_aa7a_4038_8245_77e79adbe026.slice/crio-23a505d975508ea8843d3587ada5246ccd4d155b4f536f198ee7690b634f2f85 WatchSource:0}: Error finding container 23a505d975508ea8843d3587ada5246ccd4d155b4f536f198ee7690b634f2f85: Status 404 returned error can't find the container with id 23a505d975508ea8843d3587ada5246ccd4d155b4f536f198ee7690b634f2f85 Dec 03 17:34:56 crc kubenswrapper[4787]: I1203 17:34:56.790754 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-h5xrc"] Dec 03 17:34:56 crc kubenswrapper[4787]: I1203 17:34:56.818202 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.002303 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6bd0-account-create-update-qvtnl" event={"ID":"ad8fa82a-ce58-4713-900e-f0e333cd8448","Type":"ContainerStarted","Data":"cb3a534feef18091fb89e3a046672d4271bc8f326b0529d7a8853c8063110df5"} Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.005097 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8ee8-account-create-update-bvmfg" event={"ID":"2131a38e-a365-46fd-8b7f-b01906edec67","Type":"ContainerStarted","Data":"85a879c3cdde78dee6e1f2a07b37049b87a34640339f0d4200c9e16aef4c9f9b"} Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.006333 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a415-account-create-update-t8r4c" event={"ID":"6c6d0805-aa7a-4038-8245-77e79adbe026","Type":"ContainerStarted","Data":"23a505d975508ea8843d3587ada5246ccd4d155b4f536f198ee7690b634f2f85"} Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.008339 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-mbdkp" event={"ID":"cbb0a5a5-509c-4786-8225-22ed73d9ab20","Type":"ContainerStarted","Data":"f6a6c6152806b3411274ec494179a77ff2b93e908b8635cba61275e77904fab8"} Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.017468 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4fpwj" event={"ID":"973215cc-7718-421c-b374-49e64bb9af3e","Type":"ContainerStarted","Data":"a192ed507e0ef91e9a2305e9a498ec80ea95d930f9eaad5e79f97d981206a595"} Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.019360 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"df89848b-864f-477f-bea0-e34cfcaba0fb","Type":"ContainerStarted","Data":"6e2f21296fc39e463d0c9f1080a9f3027cbf0421d13e18ba6f105c59dcdfbfe5"} Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.020914 4787 generic.go:334] "Generic (PLEG): container finished" podID="1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" containerID="d97d783c9388573039c4b54678a3dda2f9d8a8a128a5fbdef4dca8d49fba2517" exitCode=0 Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.020954 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" event={"ID":"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf","Type":"ContainerDied","Data":"d97d783c9388573039c4b54678a3dda2f9d8a8a128a5fbdef4dca8d49fba2517"} Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.033710 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-h5xrc" event={"ID":"d3bccb94-fcb5-44fc-819b-e39562a7e57a","Type":"ContainerStarted","Data":"3353f13fcc9470381af0e97bd64d65c9fe5d8717aa06ceb3324ca18f09409a5d"} Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.038536 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-mbdkp" podStartSLOduration=10.038517429 podStartE2EDuration="10.038517429s" podCreationTimestamp="2025-12-03 17:34:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:34:57.023046298 +0000 UTC m=+1313.840517247" watchObservedRunningTime="2025-12-03 17:34:57.038517429 +0000 UTC m=+1313.855988388" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.043937 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-4fpwj" podStartSLOduration=2.441881975 podStartE2EDuration="13.043915722s" podCreationTimestamp="2025-12-03 17:34:44 +0000 UTC" firstStartedPulling="2025-12-03 17:34:45.493417186 +0000 UTC m=+1302.310888145" lastFinishedPulling="2025-12-03 17:34:56.095450933 +0000 UTC m=+1312.912921892" observedRunningTime="2025-12-03 17:34:57.040967664 +0000 UTC m=+1313.858438623" watchObservedRunningTime="2025-12-03 17:34:57.043915722 +0000 UTC m=+1313.861386691" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.101722 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-fwg8j"] Dec 03 17:34:57 crc kubenswrapper[4787]: E1203 17:34:57.102846 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8dd3942-d40c-4699-8603-74f41dd659be" containerName="init" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.102873 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8dd3942-d40c-4699-8603-74f41dd659be" containerName="init" Dec 03 17:34:57 crc kubenswrapper[4787]: E1203 17:34:57.102894 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8dd3942-d40c-4699-8603-74f41dd659be" containerName="dnsmasq-dns" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.102902 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8dd3942-d40c-4699-8603-74f41dd659be" containerName="dnsmasq-dns" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.103174 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8dd3942-d40c-4699-8603-74f41dd659be" containerName="dnsmasq-dns" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.103998 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fwg8j" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.123501 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-fwg8j"] Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.203930 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73fb0e8b-a595-41f9-b732-7c0984f614e6-operator-scripts\") pod \"keystone-db-create-fwg8j\" (UID: \"73fb0e8b-a595-41f9-b732-7c0984f614e6\") " pod="openstack/keystone-db-create-fwg8j" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.204270 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2p9q\" (UniqueName: \"kubernetes.io/projected/73fb0e8b-a595-41f9-b732-7c0984f614e6-kube-api-access-q2p9q\") pod \"keystone-db-create-fwg8j\" (UID: \"73fb0e8b-a595-41f9-b732-7c0984f614e6\") " pod="openstack/keystone-db-create-fwg8j" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.207214 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kspp6"] Dec 03 17:34:57 crc kubenswrapper[4787]: W1203 17:34:57.207411 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd0bc011b_984e_4e56_a3c2_b5ea04a74a73.slice/crio-758be19b486b87d031a3e0e6a590e33442a66ee4a068cd0d71b9442c04030e65 WatchSource:0}: Error finding container 758be19b486b87d031a3e0e6a590e33442a66ee4a068cd0d71b9442c04030e65: Status 404 returned error can't find the container with id 758be19b486b87d031a3e0e6a590e33442a66ee4a068cd0d71b9442c04030e65 Dec 03 17:34:57 crc kubenswrapper[4787]: W1203 17:34:57.227151 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod40edf4b6_dc92_4d34_bbb2_a3784cf06405.slice/crio-7d0bf81931a7327aa0d4add940e0e9df18c67315b2b14becb30eb83727448ede WatchSource:0}: Error finding container 7d0bf81931a7327aa0d4add940e0e9df18c67315b2b14becb30eb83727448ede: Status 404 returned error can't find the container with id 7d0bf81931a7327aa0d4add940e0e9df18c67315b2b14becb30eb83727448ede Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.231160 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-z6tr4"] Dec 03 17:34:57 crc kubenswrapper[4787]: W1203 17:34:57.235365 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb521dcb_0316_4363_983f_1e2800ee2630.slice/crio-7626cbe01b413ecce734f27d78754e13274b25fd3790acc8b96f81ade51908f0 WatchSource:0}: Error finding container 7626cbe01b413ecce734f27d78754e13274b25fd3790acc8b96f81ade51908f0: Status 404 returned error can't find the container with id 7626cbe01b413ecce734f27d78754e13274b25fd3790acc8b96f81ade51908f0 Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.258258 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-xh2lw"] Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.278961 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-acb9-account-create-update-wv7cq"] Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.305679 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73fb0e8b-a595-41f9-b732-7c0984f614e6-operator-scripts\") pod \"keystone-db-create-fwg8j\" (UID: \"73fb0e8b-a595-41f9-b732-7c0984f614e6\") " pod="openstack/keystone-db-create-fwg8j" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.305734 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2p9q\" (UniqueName: \"kubernetes.io/projected/73fb0e8b-a595-41f9-b732-7c0984f614e6-kube-api-access-q2p9q\") pod \"keystone-db-create-fwg8j\" (UID: \"73fb0e8b-a595-41f9-b732-7c0984f614e6\") " pod="openstack/keystone-db-create-fwg8j" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.306427 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73fb0e8b-a595-41f9-b732-7c0984f614e6-operator-scripts\") pod \"keystone-db-create-fwg8j\" (UID: \"73fb0e8b-a595-41f9-b732-7c0984f614e6\") " pod="openstack/keystone-db-create-fwg8j" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.335284 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2p9q\" (UniqueName: \"kubernetes.io/projected/73fb0e8b-a595-41f9-b732-7c0984f614e6-kube-api-access-q2p9q\") pod \"keystone-db-create-fwg8j\" (UID: \"73fb0e8b-a595-41f9-b732-7c0984f614e6\") " pod="openstack/keystone-db-create-fwg8j" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.499194 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fwg8j" Dec 03 17:34:57 crc kubenswrapper[4787]: I1203 17:34:57.979866 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-fwg8j"] Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.045404 4787 generic.go:334] "Generic (PLEG): container finished" podID="d0bc011b-984e-4e56-a3c2-b5ea04a74a73" containerID="78cf7603e71e6836acd703c0ceb8e94dbaa3c540d9cd1520f9b83839afd59c3c" exitCode=0 Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.046220 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kspp6" event={"ID":"d0bc011b-984e-4e56-a3c2-b5ea04a74a73","Type":"ContainerDied","Data":"78cf7603e71e6836acd703c0ceb8e94dbaa3c540d9cd1520f9b83839afd59c3c"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.046254 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kspp6" event={"ID":"d0bc011b-984e-4e56-a3c2-b5ea04a74a73","Type":"ContainerStarted","Data":"758be19b486b87d031a3e0e6a590e33442a66ee4a068cd0d71b9442c04030e65"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.049679 4787 generic.go:334] "Generic (PLEG): container finished" podID="2131a38e-a365-46fd-8b7f-b01906edec67" containerID="93248ed1ec7fb5c91bd3039d139a9b291650631c029183d0fca52f09a2c128fa" exitCode=0 Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.049844 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8ee8-account-create-update-bvmfg" event={"ID":"2131a38e-a365-46fd-8b7f-b01906edec67","Type":"ContainerDied","Data":"93248ed1ec7fb5c91bd3039d139a9b291650631c029183d0fca52f09a2c128fa"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.055503 4787 generic.go:334] "Generic (PLEG): container finished" podID="d3bccb94-fcb5-44fc-819b-e39562a7e57a" containerID="58d22e695149a0fb7682e4f032435ed3494a52daba772e2c13d278ccb01a0466" exitCode=0 Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.055556 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-h5xrc" event={"ID":"d3bccb94-fcb5-44fc-819b-e39562a7e57a","Type":"ContainerDied","Data":"58d22e695149a0fb7682e4f032435ed3494a52daba772e2c13d278ccb01a0466"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.057078 4787 generic.go:334] "Generic (PLEG): container finished" podID="6c6d0805-aa7a-4038-8245-77e79adbe026" containerID="b0d33df450a3d278beed7b67c25c2b638db42b0e4b8191ae8d5f4335dfc63e14" exitCode=0 Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.057135 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a415-account-create-update-t8r4c" event={"ID":"6c6d0805-aa7a-4038-8245-77e79adbe026","Type":"ContainerDied","Data":"b0d33df450a3d278beed7b67c25c2b638db42b0e4b8191ae8d5f4335dfc63e14"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.060860 4787 generic.go:334] "Generic (PLEG): container finished" podID="f041eda1-f785-472b-928e-7ab2b901e99e" containerID="84641faf23ffc7fef426487234f9466e3374f556bd45a42a5246c36715f69d97" exitCode=0 Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.060967 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" event={"ID":"f041eda1-f785-472b-928e-7ab2b901e99e","Type":"ContainerDied","Data":"84641faf23ffc7fef426487234f9466e3374f556bd45a42a5246c36715f69d97"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.060995 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" event={"ID":"f041eda1-f785-472b-928e-7ab2b901e99e","Type":"ContainerStarted","Data":"034b5b370cba5a133b995fd11dfbb50fe09c1469594dac7538dccfbf1685ac82"} Dec 03 17:34:58 crc kubenswrapper[4787]: W1203 17:34:58.061591 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73fb0e8b_a595_41f9_b732_7c0984f614e6.slice/crio-161e8f96df3b305bc2e0163c8c736fa85dc0e64e4cf48a37720f63ddbd361d41 WatchSource:0}: Error finding container 161e8f96df3b305bc2e0163c8c736fa85dc0e64e4cf48a37720f63ddbd361d41: Status 404 returned error can't find the container with id 161e8f96df3b305bc2e0163c8c736fa85dc0e64e4cf48a37720f63ddbd361d41 Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.063182 4787 generic.go:334] "Generic (PLEG): container finished" podID="fb521dcb-0316-4363-983f-1e2800ee2630" containerID="cf2d60b2fed518e2538ab724aeb9860633b53fb0f42497d14dfeb0c62a7955c0" exitCode=0 Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.063254 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" event={"ID":"fb521dcb-0316-4363-983f-1e2800ee2630","Type":"ContainerDied","Data":"cf2d60b2fed518e2538ab724aeb9860633b53fb0f42497d14dfeb0c62a7955c0"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.063283 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" event={"ID":"fb521dcb-0316-4363-983f-1e2800ee2630","Type":"ContainerStarted","Data":"7626cbe01b413ecce734f27d78754e13274b25fd3790acc8b96f81ade51908f0"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.069364 4787 generic.go:334] "Generic (PLEG): container finished" podID="ad8fa82a-ce58-4713-900e-f0e333cd8448" containerID="a6379aca724ec8746c8ea782a4bbfb03129c8da4126ccbedb7ff77b67336416e" exitCode=0 Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.069564 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6bd0-account-create-update-qvtnl" event={"ID":"ad8fa82a-ce58-4713-900e-f0e333cd8448","Type":"ContainerDied","Data":"a6379aca724ec8746c8ea782a4bbfb03129c8da4126ccbedb7ff77b67336416e"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.076820 4787 generic.go:334] "Generic (PLEG): container finished" podID="40edf4b6-dc92-4d34-bbb2-a3784cf06405" containerID="4e177c0d754e8bb4b61123598ea157a8cf8c780e1d967e48235b4d539e011a3e" exitCode=0 Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.076881 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-z6tr4" event={"ID":"40edf4b6-dc92-4d34-bbb2-a3784cf06405","Type":"ContainerDied","Data":"4e177c0d754e8bb4b61123598ea157a8cf8c780e1d967e48235b4d539e011a3e"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.076904 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-z6tr4" event={"ID":"40edf4b6-dc92-4d34-bbb2-a3784cf06405","Type":"ContainerStarted","Data":"7d0bf81931a7327aa0d4add940e0e9df18c67315b2b14becb30eb83727448ede"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.078903 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" event={"ID":"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf","Type":"ContainerStarted","Data":"39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7"} Dec 03 17:34:58 crc kubenswrapper[4787]: I1203 17:34:58.196636 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" podStartSLOduration=11.196616268 podStartE2EDuration="11.196616268s" podCreationTimestamp="2025-12-03 17:34:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:34:58.188346018 +0000 UTC m=+1315.005816977" watchObservedRunningTime="2025-12-03 17:34:58.196616268 +0000 UTC m=+1315.014087227" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.092553 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"df89848b-864f-477f-bea0-e34cfcaba0fb","Type":"ContainerStarted","Data":"1dfe1e7f0bbacbdfcd8905b7360416d5c04bb0b7ba52e61af0a93c7cb685524d"} Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.092940 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"df89848b-864f-477f-bea0-e34cfcaba0fb","Type":"ContainerStarted","Data":"e40a07147d9c0efa77ddc07bc100b7e97d5dc3dbacb21cd6fc4f397d34496c1a"} Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.093064 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.095983 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kspp6" event={"ID":"d0bc011b-984e-4e56-a3c2-b5ea04a74a73","Type":"ContainerStarted","Data":"783be89bd64ad820febf63d687389dcb5e50ed3b1d5295a09290c8de26b99293"} Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.096096 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.106058 4787 generic.go:334] "Generic (PLEG): container finished" podID="73fb0e8b-a595-41f9-b732-7c0984f614e6" containerID="d92ea270829886b26a447aca6fe702d5b2d50b1c12d2e00cc73679bb7ca315ae" exitCode=0 Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.106258 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fwg8j" event={"ID":"73fb0e8b-a595-41f9-b732-7c0984f614e6","Type":"ContainerDied","Data":"d92ea270829886b26a447aca6fe702d5b2d50b1c12d2e00cc73679bb7ca315ae"} Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.106333 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fwg8j" event={"ID":"73fb0e8b-a595-41f9-b732-7c0984f614e6","Type":"ContainerStarted","Data":"161e8f96df3b305bc2e0163c8c736fa85dc0e64e4cf48a37720f63ddbd361d41"} Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.108230 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.119302 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=10.779252107 podStartE2EDuration="12.11928177s" podCreationTimestamp="2025-12-03 17:34:47 +0000 UTC" firstStartedPulling="2025-12-03 17:34:56.831914519 +0000 UTC m=+1313.649385478" lastFinishedPulling="2025-12-03 17:34:58.171944182 +0000 UTC m=+1314.989415141" observedRunningTime="2025-12-03 17:34:59.107498167 +0000 UTC m=+1315.924969126" watchObservedRunningTime="2025-12-03 17:34:59.11928177 +0000 UTC m=+1315.936752729" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.130642 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-698758b865-kspp6" podStartSLOduration=12.130623512 podStartE2EDuration="12.130623512s" podCreationTimestamp="2025-12-03 17:34:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:34:59.125335381 +0000 UTC m=+1315.942806340" watchObservedRunningTime="2025-12-03 17:34:59.130623512 +0000 UTC m=+1315.948094471" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.661676 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-z6tr4" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.750690 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pr6mx\" (UniqueName: \"kubernetes.io/projected/40edf4b6-dc92-4d34-bbb2-a3784cf06405-kube-api-access-pr6mx\") pod \"40edf4b6-dc92-4d34-bbb2-a3784cf06405\" (UID: \"40edf4b6-dc92-4d34-bbb2-a3784cf06405\") " Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.750789 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40edf4b6-dc92-4d34-bbb2-a3784cf06405-operator-scripts\") pod \"40edf4b6-dc92-4d34-bbb2-a3784cf06405\" (UID: \"40edf4b6-dc92-4d34-bbb2-a3784cf06405\") " Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.751539 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/40edf4b6-dc92-4d34-bbb2-a3784cf06405-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "40edf4b6-dc92-4d34-bbb2-a3784cf06405" (UID: "40edf4b6-dc92-4d34-bbb2-a3784cf06405"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.798393 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/40edf4b6-dc92-4d34-bbb2-a3784cf06405-kube-api-access-pr6mx" (OuterVolumeSpecName: "kube-api-access-pr6mx") pod "40edf4b6-dc92-4d34-bbb2-a3784cf06405" (UID: "40edf4b6-dc92-4d34-bbb2-a3784cf06405"). InnerVolumeSpecName "kube-api-access-pr6mx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.854185 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/40edf4b6-dc92-4d34-bbb2-a3784cf06405-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.854224 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pr6mx\" (UniqueName: \"kubernetes.io/projected/40edf4b6-dc92-4d34-bbb2-a3784cf06405-kube-api-access-pr6mx\") on node \"crc\" DevicePath \"\"" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.911608 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6bd0-account-create-update-qvtnl" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.936561 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a415-account-create-update-t8r4c" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.940406 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.949272 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8ee8-account-create-update-bvmfg" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.961501 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gd5kx\" (UniqueName: \"kubernetes.io/projected/ad8fa82a-ce58-4713-900e-f0e333cd8448-kube-api-access-gd5kx\") pod \"ad8fa82a-ce58-4713-900e-f0e333cd8448\" (UID: \"ad8fa82a-ce58-4713-900e-f0e333cd8448\") " Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.961948 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad8fa82a-ce58-4713-900e-f0e333cd8448-operator-scripts\") pod \"ad8fa82a-ce58-4713-900e-f0e333cd8448\" (UID: \"ad8fa82a-ce58-4713-900e-f0e333cd8448\") " Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.963512 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ad8fa82a-ce58-4713-900e-f0e333cd8448-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ad8fa82a-ce58-4713-900e-f0e333cd8448" (UID: "ad8fa82a-ce58-4713-900e-f0e333cd8448"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.972680 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-h5xrc" Dec 03 17:34:59 crc kubenswrapper[4787]: I1203 17:34:59.997583 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ad8fa82a-ce58-4713-900e-f0e333cd8448-kube-api-access-gd5kx" (OuterVolumeSpecName: "kube-api-access-gd5kx") pod "ad8fa82a-ce58-4713-900e-f0e333cd8448" (UID: "ad8fa82a-ce58-4713-900e-f0e333cd8448"). InnerVolumeSpecName "kube-api-access-gd5kx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.009127 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.063881 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bccb94-fcb5-44fc-819b-e39562a7e57a-operator-scripts\") pod \"d3bccb94-fcb5-44fc-819b-e39562a7e57a\" (UID: \"d3bccb94-fcb5-44fc-819b-e39562a7e57a\") " Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.063996 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f041eda1-f785-472b-928e-7ab2b901e99e-operator-scripts\") pod \"f041eda1-f785-472b-928e-7ab2b901e99e\" (UID: \"f041eda1-f785-472b-928e-7ab2b901e99e\") " Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.064038 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb521dcb-0316-4363-983f-1e2800ee2630-operator-scripts\") pod \"fb521dcb-0316-4363-983f-1e2800ee2630\" (UID: \"fb521dcb-0316-4363-983f-1e2800ee2630\") " Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.064192 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gh54\" (UniqueName: \"kubernetes.io/projected/fb521dcb-0316-4363-983f-1e2800ee2630-kube-api-access-6gh54\") pod \"fb521dcb-0316-4363-983f-1e2800ee2630\" (UID: \"fb521dcb-0316-4363-983f-1e2800ee2630\") " Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.064271 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxszf\" (UniqueName: \"kubernetes.io/projected/d3bccb94-fcb5-44fc-819b-e39562a7e57a-kube-api-access-jxszf\") pod \"d3bccb94-fcb5-44fc-819b-e39562a7e57a\" (UID: \"d3bccb94-fcb5-44fc-819b-e39562a7e57a\") " Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.064646 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c6d0805-aa7a-4038-8245-77e79adbe026-operator-scripts\") pod \"6c6d0805-aa7a-4038-8245-77e79adbe026\" (UID: \"6c6d0805-aa7a-4038-8245-77e79adbe026\") " Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.064467 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f041eda1-f785-472b-928e-7ab2b901e99e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f041eda1-f785-472b-928e-7ab2b901e99e" (UID: "f041eda1-f785-472b-928e-7ab2b901e99e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.064687 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5spth\" (UniqueName: \"kubernetes.io/projected/2131a38e-a365-46fd-8b7f-b01906edec67-kube-api-access-5spth\") pod \"2131a38e-a365-46fd-8b7f-b01906edec67\" (UID: \"2131a38e-a365-46fd-8b7f-b01906edec67\") " Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.064575 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fb521dcb-0316-4363-983f-1e2800ee2630-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fb521dcb-0316-4363-983f-1e2800ee2630" (UID: "fb521dcb-0316-4363-983f-1e2800ee2630"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.064713 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdq78\" (UniqueName: \"kubernetes.io/projected/f041eda1-f785-472b-928e-7ab2b901e99e-kube-api-access-mdq78\") pod \"f041eda1-f785-472b-928e-7ab2b901e99e\" (UID: \"f041eda1-f785-472b-928e-7ab2b901e99e\") " Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.064740 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2131a38e-a365-46fd-8b7f-b01906edec67-operator-scripts\") pod \"2131a38e-a365-46fd-8b7f-b01906edec67\" (UID: \"2131a38e-a365-46fd-8b7f-b01906edec67\") " Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.064772 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfrnl\" (UniqueName: \"kubernetes.io/projected/6c6d0805-aa7a-4038-8245-77e79adbe026-kube-api-access-cfrnl\") pod \"6c6d0805-aa7a-4038-8245-77e79adbe026\" (UID: \"6c6d0805-aa7a-4038-8245-77e79adbe026\") " Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.065299 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f041eda1-f785-472b-928e-7ab2b901e99e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.065322 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fb521dcb-0316-4363-983f-1e2800ee2630-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.065335 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ad8fa82a-ce58-4713-900e-f0e333cd8448-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.065346 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gd5kx\" (UniqueName: \"kubernetes.io/projected/ad8fa82a-ce58-4713-900e-f0e333cd8448-kube-api-access-gd5kx\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.065280 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6c6d0805-aa7a-4038-8245-77e79adbe026-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6c6d0805-aa7a-4038-8245-77e79adbe026" (UID: "6c6d0805-aa7a-4038-8245-77e79adbe026"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.065489 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2131a38e-a365-46fd-8b7f-b01906edec67-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2131a38e-a365-46fd-8b7f-b01906edec67" (UID: "2131a38e-a365-46fd-8b7f-b01906edec67"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.065840 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3bccb94-fcb5-44fc-819b-e39562a7e57a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d3bccb94-fcb5-44fc-819b-e39562a7e57a" (UID: "d3bccb94-fcb5-44fc-819b-e39562a7e57a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.071712 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2131a38e-a365-46fd-8b7f-b01906edec67-kube-api-access-5spth" (OuterVolumeSpecName: "kube-api-access-5spth") pod "2131a38e-a365-46fd-8b7f-b01906edec67" (UID: "2131a38e-a365-46fd-8b7f-b01906edec67"). InnerVolumeSpecName "kube-api-access-5spth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.071751 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f041eda1-f785-472b-928e-7ab2b901e99e-kube-api-access-mdq78" (OuterVolumeSpecName: "kube-api-access-mdq78") pod "f041eda1-f785-472b-928e-7ab2b901e99e" (UID: "f041eda1-f785-472b-928e-7ab2b901e99e"). InnerVolumeSpecName "kube-api-access-mdq78". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.072221 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3bccb94-fcb5-44fc-819b-e39562a7e57a-kube-api-access-jxszf" (OuterVolumeSpecName: "kube-api-access-jxszf") pod "d3bccb94-fcb5-44fc-819b-e39562a7e57a" (UID: "d3bccb94-fcb5-44fc-819b-e39562a7e57a"). InnerVolumeSpecName "kube-api-access-jxszf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.072786 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb521dcb-0316-4363-983f-1e2800ee2630-kube-api-access-6gh54" (OuterVolumeSpecName: "kube-api-access-6gh54") pod "fb521dcb-0316-4363-983f-1e2800ee2630" (UID: "fb521dcb-0316-4363-983f-1e2800ee2630"). InnerVolumeSpecName "kube-api-access-6gh54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.073568 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6c6d0805-aa7a-4038-8245-77e79adbe026-kube-api-access-cfrnl" (OuterVolumeSpecName: "kube-api-access-cfrnl") pod "6c6d0805-aa7a-4038-8245-77e79adbe026" (UID: "6c6d0805-aa7a-4038-8245-77e79adbe026"). InnerVolumeSpecName "kube-api-access-cfrnl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.117347 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-h5xrc" event={"ID":"d3bccb94-fcb5-44fc-819b-e39562a7e57a","Type":"ContainerDied","Data":"3353f13fcc9470381af0e97bd64d65c9fe5d8717aa06ceb3324ca18f09409a5d"} Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.117394 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3353f13fcc9470381af0e97bd64d65c9fe5d8717aa06ceb3324ca18f09409a5d" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.117392 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-h5xrc" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.118907 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-z6tr4" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.118996 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-z6tr4" event={"ID":"40edf4b6-dc92-4d34-bbb2-a3784cf06405","Type":"ContainerDied","Data":"7d0bf81931a7327aa0d4add940e0e9df18c67315b2b14becb30eb83727448ede"} Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.119043 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d0bf81931a7327aa0d4add940e0e9df18c67315b2b14becb30eb83727448ede" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.120064 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-a415-account-create-update-t8r4c" event={"ID":"6c6d0805-aa7a-4038-8245-77e79adbe026","Type":"ContainerDied","Data":"23a505d975508ea8843d3587ada5246ccd4d155b4f536f198ee7690b634f2f85"} Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.120090 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-a415-account-create-update-t8r4c" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.120100 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="23a505d975508ea8843d3587ada5246ccd4d155b4f536f198ee7690b634f2f85" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.121131 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" event={"ID":"f041eda1-f785-472b-928e-7ab2b901e99e","Type":"ContainerDied","Data":"034b5b370cba5a133b995fd11dfbb50fe09c1469594dac7538dccfbf1685ac82"} Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.121158 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="034b5b370cba5a133b995fd11dfbb50fe09c1469594dac7538dccfbf1685ac82" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.121163 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-acb9-account-create-update-wv7cq" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.122369 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.122373 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-db-create-xh2lw" event={"ID":"fb521dcb-0316-4363-983f-1e2800ee2630","Type":"ContainerDied","Data":"7626cbe01b413ecce734f27d78754e13274b25fd3790acc8b96f81ade51908f0"} Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.122401 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7626cbe01b413ecce734f27d78754e13274b25fd3790acc8b96f81ade51908f0" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.123450 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-6bd0-account-create-update-qvtnl" event={"ID":"ad8fa82a-ce58-4713-900e-f0e333cd8448","Type":"ContainerDied","Data":"cb3a534feef18091fb89e3a046672d4271bc8f326b0529d7a8853c8063110df5"} Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.123465 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-6bd0-account-create-update-qvtnl" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.123476 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cb3a534feef18091fb89e3a046672d4271bc8f326b0529d7a8853c8063110df5" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.125132 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8ee8-account-create-update-bvmfg" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.128327 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8ee8-account-create-update-bvmfg" event={"ID":"2131a38e-a365-46fd-8b7f-b01906edec67","Type":"ContainerDied","Data":"85a879c3cdde78dee6e1f2a07b37049b87a34640339f0d4200c9e16aef4c9f9b"} Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.128363 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85a879c3cdde78dee6e1f2a07b37049b87a34640339f0d4200c9e16aef4c9f9b" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.166916 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2131a38e-a365-46fd-8b7f-b01906edec67-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.166956 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfrnl\" (UniqueName: \"kubernetes.io/projected/6c6d0805-aa7a-4038-8245-77e79adbe026-kube-api-access-cfrnl\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.166969 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d3bccb94-fcb5-44fc-819b-e39562a7e57a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.166981 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gh54\" (UniqueName: \"kubernetes.io/projected/fb521dcb-0316-4363-983f-1e2800ee2630-kube-api-access-6gh54\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.166992 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxszf\" (UniqueName: \"kubernetes.io/projected/d3bccb94-fcb5-44fc-819b-e39562a7e57a-kube-api-access-jxszf\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.167001 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6c6d0805-aa7a-4038-8245-77e79adbe026-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.167011 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5spth\" (UniqueName: \"kubernetes.io/projected/2131a38e-a365-46fd-8b7f-b01906edec67-kube-api-access-5spth\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:00 crc kubenswrapper[4787]: I1203 17:35:00.167041 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdq78\" (UniqueName: \"kubernetes.io/projected/f041eda1-f785-472b-928e-7ab2b901e99e-kube-api-access-mdq78\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:02 crc kubenswrapper[4787]: I1203 17:35:02.537195 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:35:02 crc kubenswrapper[4787]: I1203 17:35:02.956388 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fwg8j" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.022064 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73fb0e8b-a595-41f9-b732-7c0984f614e6-operator-scripts\") pod \"73fb0e8b-a595-41f9-b732-7c0984f614e6\" (UID: \"73fb0e8b-a595-41f9-b732-7c0984f614e6\") " Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.022957 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2p9q\" (UniqueName: \"kubernetes.io/projected/73fb0e8b-a595-41f9-b732-7c0984f614e6-kube-api-access-q2p9q\") pod \"73fb0e8b-a595-41f9-b732-7c0984f614e6\" (UID: \"73fb0e8b-a595-41f9-b732-7c0984f614e6\") " Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.023473 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73fb0e8b-a595-41f9-b732-7c0984f614e6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "73fb0e8b-a595-41f9-b732-7c0984f614e6" (UID: "73fb0e8b-a595-41f9-b732-7c0984f614e6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.024003 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/73fb0e8b-a595-41f9-b732-7c0984f614e6-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.038414 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73fb0e8b-a595-41f9-b732-7c0984f614e6-kube-api-access-q2p9q" (OuterVolumeSpecName: "kube-api-access-q2p9q") pod "73fb0e8b-a595-41f9-b732-7c0984f614e6" (UID: "73fb0e8b-a595-41f9-b732-7c0984f614e6"). InnerVolumeSpecName "kube-api-access-q2p9q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.053718 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-wcnd6"] Dec 03 17:35:03 crc kubenswrapper[4787]: E1203 17:35:03.054424 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3bccb94-fcb5-44fc-819b-e39562a7e57a" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054441 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3bccb94-fcb5-44fc-819b-e39562a7e57a" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: E1203 17:35:03.054460 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73fb0e8b-a595-41f9-b732-7c0984f614e6" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054467 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="73fb0e8b-a595-41f9-b732-7c0984f614e6" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: E1203 17:35:03.054483 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ad8fa82a-ce58-4713-900e-f0e333cd8448" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054490 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ad8fa82a-ce58-4713-900e-f0e333cd8448" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: E1203 17:35:03.054503 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="40edf4b6-dc92-4d34-bbb2-a3784cf06405" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054510 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="40edf4b6-dc92-4d34-bbb2-a3784cf06405" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: E1203 17:35:03.054535 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2131a38e-a365-46fd-8b7f-b01906edec67" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054543 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2131a38e-a365-46fd-8b7f-b01906edec67" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: E1203 17:35:03.054554 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb521dcb-0316-4363-983f-1e2800ee2630" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054561 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb521dcb-0316-4363-983f-1e2800ee2630" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: E1203 17:35:03.054574 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6c6d0805-aa7a-4038-8245-77e79adbe026" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054581 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6c6d0805-aa7a-4038-8245-77e79adbe026" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: E1203 17:35:03.054598 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f041eda1-f785-472b-928e-7ab2b901e99e" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054605 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f041eda1-f785-472b-928e-7ab2b901e99e" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054827 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f041eda1-f785-472b-928e-7ab2b901e99e" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054846 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ad8fa82a-ce58-4713-900e-f0e333cd8448" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054861 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb521dcb-0316-4363-983f-1e2800ee2630" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054872 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3bccb94-fcb5-44fc-819b-e39562a7e57a" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054885 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="73fb0e8b-a595-41f9-b732-7c0984f614e6" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054894 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6c6d0805-aa7a-4038-8245-77e79adbe026" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054903 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2131a38e-a365-46fd-8b7f-b01906edec67" containerName="mariadb-account-create-update" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.054910 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="40edf4b6-dc92-4d34-bbb2-a3784cf06405" containerName="mariadb-database-create" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.056140 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.059368 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.059608 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8g2xg" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.060230 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.084714 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-wcnd6"] Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.125365 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqpwk\" (UniqueName: \"kubernetes.io/projected/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-kube-api-access-nqpwk\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.125452 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-combined-ca-bundle\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.125536 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-db-sync-config-data\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.125577 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-config-data\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.125695 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2p9q\" (UniqueName: \"kubernetes.io/projected/73fb0e8b-a595-41f9-b732-7c0984f614e6-kube-api-access-q2p9q\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.172951 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-7kjsl"] Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.236660 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqpwk\" (UniqueName: \"kubernetes.io/projected/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-kube-api-access-nqpwk\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.236774 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-combined-ca-bundle\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.236916 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-db-sync-config-data\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.236970 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-config-data\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.238140 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" podUID="1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" containerName="dnsmasq-dns" containerID="cri-o://39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7" gracePeriod=10 Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.238182 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-fwg8j" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.238282 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-fwg8j" event={"ID":"73fb0e8b-a595-41f9-b732-7c0984f614e6","Type":"ContainerDied","Data":"161e8f96df3b305bc2e0163c8c736fa85dc0e64e4cf48a37720f63ddbd361d41"} Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.238341 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="161e8f96df3b305bc2e0163c8c736fa85dc0e64e4cf48a37720f63ddbd361d41" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.274090 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-db-sync-config-data\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.274312 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-combined-ca-bundle\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.279635 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqpwk\" (UniqueName: \"kubernetes.io/projected/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-kube-api-access-nqpwk\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.282158 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-config-data\") pod \"glance-db-sync-wcnd6\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.417815 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.842117 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.967248 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qnrbz\" (UniqueName: \"kubernetes.io/projected/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-kube-api-access-qnrbz\") pod \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.967710 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-dns-svc\") pod \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.967797 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-config\") pod \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.967826 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-ovsdbserver-sb\") pod \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\" (UID: \"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf\") " Dec 03 17:35:03 crc kubenswrapper[4787]: I1203 17:35:03.976015 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-kube-api-access-qnrbz" (OuterVolumeSpecName: "kube-api-access-qnrbz") pod "1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" (UID: "1c97c1be-49ce-4ab1-b3aa-f4055032cfdf"). InnerVolumeSpecName "kube-api-access-qnrbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.069298 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" (UID: "1c97c1be-49ce-4ab1-b3aa-f4055032cfdf"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.072755 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qnrbz\" (UniqueName: \"kubernetes.io/projected/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-kube-api-access-qnrbz\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.072799 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.087190 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" (UID: "1c97c1be-49ce-4ab1-b3aa-f4055032cfdf"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.094860 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-config" (OuterVolumeSpecName: "config") pod "1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" (UID: "1c97c1be-49ce-4ab1-b3aa-f4055032cfdf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.147965 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-wcnd6"] Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.175348 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.175626 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.259008 4787 generic.go:334] "Generic (PLEG): container finished" podID="1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" containerID="39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7" exitCode=0 Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.259145 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" event={"ID":"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf","Type":"ContainerDied","Data":"39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7"} Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.259198 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" event={"ID":"1c97c1be-49ce-4ab1-b3aa-f4055032cfdf","Type":"ContainerDied","Data":"f4fa2a31d0a4bcdf092be33a1bf0cbd3126fe1e1fe9ff86c111f873f08c66efb"} Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.259231 4787 scope.go:117] "RemoveContainer" containerID="39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.259456 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6c89d5d749-7kjsl" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.284254 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3b68d021-5db8-4a87-92f2-e827bbdd2255","Type":"ContainerStarted","Data":"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb"} Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.286214 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wcnd6" event={"ID":"948f9158-3e6d-4a5d-a8d4-1c9480a85f90","Type":"ContainerStarted","Data":"ae5f0c500fe7b34fd2562552c411ca64313dc46aabeb746c19bb606ee9b520bc"} Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.303401 4787 scope.go:117] "RemoveContainer" containerID="d97d783c9388573039c4b54678a3dda2f9d8a8a128a5fbdef4dca8d49fba2517" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.307769 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-7kjsl"] Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.317831 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6c89d5d749-7kjsl"] Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.343197 4787 scope.go:117] "RemoveContainer" containerID="39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7" Dec 03 17:35:04 crc kubenswrapper[4787]: E1203 17:35:04.344092 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7\": container with ID starting with 39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7 not found: ID does not exist" containerID="39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.344147 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7"} err="failed to get container status \"39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7\": rpc error: code = NotFound desc = could not find container \"39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7\": container with ID starting with 39ae8222f4e1f910d3692a923d086e5fc6dd8bd1d66e9fa9bd152abfe3d160e7 not found: ID does not exist" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.344181 4787 scope.go:117] "RemoveContainer" containerID="d97d783c9388573039c4b54678a3dda2f9d8a8a128a5fbdef4dca8d49fba2517" Dec 03 17:35:04 crc kubenswrapper[4787]: E1203 17:35:04.344851 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d97d783c9388573039c4b54678a3dda2f9d8a8a128a5fbdef4dca8d49fba2517\": container with ID starting with d97d783c9388573039c4b54678a3dda2f9d8a8a128a5fbdef4dca8d49fba2517 not found: ID does not exist" containerID="d97d783c9388573039c4b54678a3dda2f9d8a8a128a5fbdef4dca8d49fba2517" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.344933 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d97d783c9388573039c4b54678a3dda2f9d8a8a128a5fbdef4dca8d49fba2517"} err="failed to get container status \"d97d783c9388573039c4b54678a3dda2f9d8a8a128a5fbdef4dca8d49fba2517\": rpc error: code = NotFound desc = could not find container \"d97d783c9388573039c4b54678a3dda2f9d8a8a128a5fbdef4dca8d49fba2517\": container with ID starting with d97d783c9388573039c4b54678a3dda2f9d8a8a128a5fbdef4dca8d49fba2517 not found: ID does not exist" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.714637 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj"] Dec 03 17:35:04 crc kubenswrapper[4787]: E1203 17:35:04.715601 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" containerName="dnsmasq-dns" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.715624 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" containerName="dnsmasq-dns" Dec 03 17:35:04 crc kubenswrapper[4787]: E1203 17:35:04.715662 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" containerName="init" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.715669 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" containerName="init" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.715856 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" containerName="dnsmasq-dns" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.716770 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.751561 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj"] Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.808996 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ec64d7f-0783-4048-a8ac-baa6de52ed78-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-dhhvj\" (UID: \"4ec64d7f-0783-4048-a8ac-baa6de52ed78\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.809087 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxvwl\" (UniqueName: \"kubernetes.io/projected/4ec64d7f-0783-4048-a8ac-baa6de52ed78-kube-api-access-gxvwl\") pod \"mysqld-exporter-openstack-cell1-db-create-dhhvj\" (UID: \"4ec64d7f-0783-4048-a8ac-baa6de52ed78\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.910783 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ec64d7f-0783-4048-a8ac-baa6de52ed78-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-dhhvj\" (UID: \"4ec64d7f-0783-4048-a8ac-baa6de52ed78\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.910867 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxvwl\" (UniqueName: \"kubernetes.io/projected/4ec64d7f-0783-4048-a8ac-baa6de52ed78-kube-api-access-gxvwl\") pod \"mysqld-exporter-openstack-cell1-db-create-dhhvj\" (UID: \"4ec64d7f-0783-4048-a8ac-baa6de52ed78\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" Dec 03 17:35:04 crc kubenswrapper[4787]: I1203 17:35:04.911661 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ec64d7f-0783-4048-a8ac-baa6de52ed78-operator-scripts\") pod \"mysqld-exporter-openstack-cell1-db-create-dhhvj\" (UID: \"4ec64d7f-0783-4048-a8ac-baa6de52ed78\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.099697 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-960e-account-create-update-8dsds"] Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.101547 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.107616 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-openstack-cell1-db-secret" Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.109410 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxvwl\" (UniqueName: \"kubernetes.io/projected/4ec64d7f-0783-4048-a8ac-baa6de52ed78-kube-api-access-gxvwl\") pod \"mysqld-exporter-openstack-cell1-db-create-dhhvj\" (UID: \"4ec64d7f-0783-4048-a8ac-baa6de52ed78\") " pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.123700 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-960e-account-create-update-8dsds"] Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.226507 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scz6p\" (UniqueName: \"kubernetes.io/projected/c3b80bcb-cab7-4a0e-a761-ad5942435422-kube-api-access-scz6p\") pod \"mysqld-exporter-960e-account-create-update-8dsds\" (UID: \"c3b80bcb-cab7-4a0e-a761-ad5942435422\") " pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.226564 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3b80bcb-cab7-4a0e-a761-ad5942435422-operator-scripts\") pod \"mysqld-exporter-960e-account-create-update-8dsds\" (UID: \"c3b80bcb-cab7-4a0e-a761-ad5942435422\") " pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.305344 4787 generic.go:334] "Generic (PLEG): container finished" podID="973215cc-7718-421c-b374-49e64bb9af3e" containerID="a192ed507e0ef91e9a2305e9a498ec80ea95d930f9eaad5e79f97d981206a595" exitCode=0 Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.305398 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4fpwj" event={"ID":"973215cc-7718-421c-b374-49e64bb9af3e","Type":"ContainerDied","Data":"a192ed507e0ef91e9a2305e9a498ec80ea95d930f9eaad5e79f97d981206a595"} Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.337515 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scz6p\" (UniqueName: \"kubernetes.io/projected/c3b80bcb-cab7-4a0e-a761-ad5942435422-kube-api-access-scz6p\") pod \"mysqld-exporter-960e-account-create-update-8dsds\" (UID: \"c3b80bcb-cab7-4a0e-a761-ad5942435422\") " pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.337598 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3b80bcb-cab7-4a0e-a761-ad5942435422-operator-scripts\") pod \"mysqld-exporter-960e-account-create-update-8dsds\" (UID: \"c3b80bcb-cab7-4a0e-a761-ad5942435422\") " pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.338533 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3b80bcb-cab7-4a0e-a761-ad5942435422-operator-scripts\") pod \"mysqld-exporter-960e-account-create-update-8dsds\" (UID: \"c3b80bcb-cab7-4a0e-a761-ad5942435422\") " pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.355706 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scz6p\" (UniqueName: \"kubernetes.io/projected/c3b80bcb-cab7-4a0e-a761-ad5942435422-kube-api-access-scz6p\") pod \"mysqld-exporter-960e-account-create-update-8dsds\" (UID: \"c3b80bcb-cab7-4a0e-a761-ad5942435422\") " pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.388578 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" Dec 03 17:35:05 crc kubenswrapper[4787]: I1203 17:35:05.526004 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:05.800549 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c97c1be-49ce-4ab1-b3aa-f4055032cfdf" path="/var/lib/kubelet/pods/1c97c1be-49ce-4ab1-b3aa-f4055032cfdf/volumes" Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.890295 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-85474b5df5-nrlj4" podUID="e4903b58-00fb-44e2-a2a1-60f89894b73c" containerName="console" containerID="cri-o://9c805ac3f15f3ac1131c700a4c872396df24d2d62ff0e5da005e0ec0c8e6132b" gracePeriod=15 Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.892107 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.925773 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj"] Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.949891 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-960e-account-create-update-8dsds"] Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.973885 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-ring-data-devices\") pod \"973215cc-7718-421c-b374-49e64bb9af3e\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.973961 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-swiftconf\") pod \"973215cc-7718-421c-b374-49e64bb9af3e\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.974049 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-combined-ca-bundle\") pod \"973215cc-7718-421c-b374-49e64bb9af3e\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.974090 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-dispersionconf\") pod \"973215cc-7718-421c-b374-49e64bb9af3e\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.974158 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-scripts\") pod \"973215cc-7718-421c-b374-49e64bb9af3e\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.974182 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n2wts\" (UniqueName: \"kubernetes.io/projected/973215cc-7718-421c-b374-49e64bb9af3e-kube-api-access-n2wts\") pod \"973215cc-7718-421c-b374-49e64bb9af3e\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.974264 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/973215cc-7718-421c-b374-49e64bb9af3e-etc-swift\") pod \"973215cc-7718-421c-b374-49e64bb9af3e\" (UID: \"973215cc-7718-421c-b374-49e64bb9af3e\") " Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.975017 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "973215cc-7718-421c-b374-49e64bb9af3e" (UID: "973215cc-7718-421c-b374-49e64bb9af3e"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.975313 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/973215cc-7718-421c-b374-49e64bb9af3e-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "973215cc-7718-421c-b374-49e64bb9af3e" (UID: "973215cc-7718-421c-b374-49e64bb9af3e"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.981319 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/973215cc-7718-421c-b374-49e64bb9af3e-kube-api-access-n2wts" (OuterVolumeSpecName: "kube-api-access-n2wts") pod "973215cc-7718-421c-b374-49e64bb9af3e" (UID: "973215cc-7718-421c-b374-49e64bb9af3e"). InnerVolumeSpecName "kube-api-access-n2wts". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:06 crc kubenswrapper[4787]: I1203 17:35:06.984692 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "973215cc-7718-421c-b374-49e64bb9af3e" (UID: "973215cc-7718-421c-b374-49e64bb9af3e"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:06 crc kubenswrapper[4787]: W1203 17:35:06.984886 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ec64d7f_0783_4048_a8ac_baa6de52ed78.slice/crio-a032b1313c52f1d5b2f63d146ea40e30f79eb8076a9cbf38f95f8ae0f92e2b94 WatchSource:0}: Error finding container a032b1313c52f1d5b2f63d146ea40e30f79eb8076a9cbf38f95f8ae0f92e2b94: Status 404 returned error can't find the container with id a032b1313c52f1d5b2f63d146ea40e30f79eb8076a9cbf38f95f8ae0f92e2b94 Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.002722 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "973215cc-7718-421c-b374-49e64bb9af3e" (UID: "973215cc-7718-421c-b374-49e64bb9af3e"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.004692 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "973215cc-7718-421c-b374-49e64bb9af3e" (UID: "973215cc-7718-421c-b374-49e64bb9af3e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.026542 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-scripts" (OuterVolumeSpecName: "scripts") pod "973215cc-7718-421c-b374-49e64bb9af3e" (UID: "973215cc-7718-421c-b374-49e64bb9af3e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.076007 4787 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-swiftconf\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.076053 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.076065 4787 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/973215cc-7718-421c-b374-49e64bb9af3e-dispersionconf\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.076073 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.076082 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n2wts\" (UniqueName: \"kubernetes.io/projected/973215cc-7718-421c-b374-49e64bb9af3e-kube-api-access-n2wts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.076093 4787 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/973215cc-7718-421c-b374-49e64bb9af3e-etc-swift\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.076103 4787 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/973215cc-7718-421c-b374-49e64bb9af3e-ring-data-devices\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.339679 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" event={"ID":"c3b80bcb-cab7-4a0e-a761-ad5942435422","Type":"ContainerStarted","Data":"31c94fc96a99da5e5eb901fcd8b315d2db0356c0e5377be64b601229f62c5736"} Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.340098 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" event={"ID":"c3b80bcb-cab7-4a0e-a761-ad5942435422","Type":"ContainerStarted","Data":"4116a2ec5d4e3cc7f4e9634604496484d1daed2044848dc5cf6a5b0ed52629f5"} Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.353371 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-4fpwj" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.353385 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-4fpwj" event={"ID":"973215cc-7718-421c-b374-49e64bb9af3e","Type":"ContainerDied","Data":"0a469acb7fca797ef599a0e1917fb95f46bef8eb8c15499bcb29111d2e56e19a"} Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.353492 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a469acb7fca797ef599a0e1917fb95f46bef8eb8c15499bcb29111d2e56e19a" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.364874 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" event={"ID":"4ec64d7f-0783-4048-a8ac-baa6de52ed78","Type":"ContainerStarted","Data":"9a8da11cb9d3a4532da1b97a323c979a02e6f2fbd5e26b282773c81d9a63d9af"} Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.364923 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" event={"ID":"4ec64d7f-0783-4048-a8ac-baa6de52ed78","Type":"ContainerStarted","Data":"a032b1313c52f1d5b2f63d146ea40e30f79eb8076a9cbf38f95f8ae0f92e2b94"} Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.372187 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3b68d021-5db8-4a87-92f2-e827bbdd2255","Type":"ContainerStarted","Data":"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213"} Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.374477 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-85474b5df5-nrlj4_e4903b58-00fb-44e2-a2a1-60f89894b73c/console/0.log" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.374510 4787 generic.go:334] "Generic (PLEG): container finished" podID="e4903b58-00fb-44e2-a2a1-60f89894b73c" containerID="9c805ac3f15f3ac1131c700a4c872396df24d2d62ff0e5da005e0ec0c8e6132b" exitCode=2 Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.374546 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-85474b5df5-nrlj4" event={"ID":"e4903b58-00fb-44e2-a2a1-60f89894b73c","Type":"ContainerDied","Data":"9c805ac3f15f3ac1131c700a4c872396df24d2d62ff0e5da005e0ec0c8e6132b"} Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.377351 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" podStartSLOduration=2.377329512 podStartE2EDuration="2.377329512s" podCreationTimestamp="2025-12-03 17:35:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:35:07.362262762 +0000 UTC m=+1324.179733721" watchObservedRunningTime="2025-12-03 17:35:07.377329512 +0000 UTC m=+1324.194800471" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.388795 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" podStartSLOduration=3.388776356 podStartE2EDuration="3.388776356s" podCreationTimestamp="2025-12-03 17:35:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:35:07.387335558 +0000 UTC m=+1324.204806517" watchObservedRunningTime="2025-12-03 17:35:07.388776356 +0000 UTC m=+1324.206247325" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.423942 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-85474b5df5-nrlj4_e4903b58-00fb-44e2-a2a1-60f89894b73c/console/0.log" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.424064 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.502408 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-oauth-config\") pod \"e4903b58-00fb-44e2-a2a1-60f89894b73c\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.503837 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-oauth-serving-cert\") pod \"e4903b58-00fb-44e2-a2a1-60f89894b73c\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.505192 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-config\") pod \"e4903b58-00fb-44e2-a2a1-60f89894b73c\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.504620 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "e4903b58-00fb-44e2-a2a1-60f89894b73c" (UID: "e4903b58-00fb-44e2-a2a1-60f89894b73c"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.505250 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kthzp\" (UniqueName: \"kubernetes.io/projected/e4903b58-00fb-44e2-a2a1-60f89894b73c-kube-api-access-kthzp\") pod \"e4903b58-00fb-44e2-a2a1-60f89894b73c\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.506063 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-service-ca\") pod \"e4903b58-00fb-44e2-a2a1-60f89894b73c\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.506202 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-trusted-ca-bundle\") pod \"e4903b58-00fb-44e2-a2a1-60f89894b73c\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.506460 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-serving-cert\") pod \"e4903b58-00fb-44e2-a2a1-60f89894b73c\" (UID: \"e4903b58-00fb-44e2-a2a1-60f89894b73c\") " Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.507058 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-service-ca" (OuterVolumeSpecName: "service-ca") pod "e4903b58-00fb-44e2-a2a1-60f89894b73c" (UID: "e4903b58-00fb-44e2-a2a1-60f89894b73c"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.507328 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-config" (OuterVolumeSpecName: "console-config") pod "e4903b58-00fb-44e2-a2a1-60f89894b73c" (UID: "e4903b58-00fb-44e2-a2a1-60f89894b73c"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.507898 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "e4903b58-00fb-44e2-a2a1-60f89894b73c" (UID: "e4903b58-00fb-44e2-a2a1-60f89894b73c"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.509232 4787 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.509274 4787 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.509288 4787 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-service-ca\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.509299 4787 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e4903b58-00fb-44e2-a2a1-60f89894b73c-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.511361 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "e4903b58-00fb-44e2-a2a1-60f89894b73c" (UID: "e4903b58-00fb-44e2-a2a1-60f89894b73c"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.511814 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "e4903b58-00fb-44e2-a2a1-60f89894b73c" (UID: "e4903b58-00fb-44e2-a2a1-60f89894b73c"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.515008 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e4903b58-00fb-44e2-a2a1-60f89894b73c-kube-api-access-kthzp" (OuterVolumeSpecName: "kube-api-access-kthzp") pod "e4903b58-00fb-44e2-a2a1-60f89894b73c" (UID: "e4903b58-00fb-44e2-a2a1-60f89894b73c"). InnerVolumeSpecName "kube-api-access-kthzp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.611196 4787 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.611237 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kthzp\" (UniqueName: \"kubernetes.io/projected/e4903b58-00fb-44e2-a2a1-60f89894b73c-kube-api-access-kthzp\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:07 crc kubenswrapper[4787]: I1203 17:35:07.611248 4787 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/e4903b58-00fb-44e2-a2a1-60f89894b73c-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:08 crc kubenswrapper[4787]: I1203 17:35:08.384189 4787 generic.go:334] "Generic (PLEG): container finished" podID="4ec64d7f-0783-4048-a8ac-baa6de52ed78" containerID="9a8da11cb9d3a4532da1b97a323c979a02e6f2fbd5e26b282773c81d9a63d9af" exitCode=0 Dec 03 17:35:08 crc kubenswrapper[4787]: I1203 17:35:08.384514 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" event={"ID":"4ec64d7f-0783-4048-a8ac-baa6de52ed78","Type":"ContainerDied","Data":"9a8da11cb9d3a4532da1b97a323c979a02e6f2fbd5e26b282773c81d9a63d9af"} Dec 03 17:35:08 crc kubenswrapper[4787]: I1203 17:35:08.387171 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-85474b5df5-nrlj4" event={"ID":"e4903b58-00fb-44e2-a2a1-60f89894b73c","Type":"ContainerDied","Data":"b75e6512fd9f5ac4f775f122247c3548c9f7604088aca906209ccd78ae9cbea0"} Dec 03 17:35:08 crc kubenswrapper[4787]: I1203 17:35:08.387207 4787 scope.go:117] "RemoveContainer" containerID="9c805ac3f15f3ac1131c700a4c872396df24d2d62ff0e5da005e0ec0c8e6132b" Dec 03 17:35:08 crc kubenswrapper[4787]: I1203 17:35:08.387295 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-85474b5df5-nrlj4" Dec 03 17:35:08 crc kubenswrapper[4787]: I1203 17:35:08.389708 4787 generic.go:334] "Generic (PLEG): container finished" podID="c3b80bcb-cab7-4a0e-a761-ad5942435422" containerID="31c94fc96a99da5e5eb901fcd8b315d2db0356c0e5377be64b601229f62c5736" exitCode=0 Dec 03 17:35:08 crc kubenswrapper[4787]: I1203 17:35:08.389742 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" event={"ID":"c3b80bcb-cab7-4a0e-a761-ad5942435422","Type":"ContainerDied","Data":"31c94fc96a99da5e5eb901fcd8b315d2db0356c0e5377be64b601229f62c5736"} Dec 03 17:35:08 crc kubenswrapper[4787]: I1203 17:35:08.449540 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-85474b5df5-nrlj4"] Dec 03 17:35:08 crc kubenswrapper[4787]: I1203 17:35:08.458806 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-85474b5df5-nrlj4"] Dec 03 17:35:09 crc kubenswrapper[4787]: I1203 17:35:09.779438 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e4903b58-00fb-44e2-a2a1-60f89894b73c" path="/var/lib/kubelet/pods/e4903b58-00fb-44e2-a2a1-60f89894b73c/volumes" Dec 03 17:35:09 crc kubenswrapper[4787]: I1203 17:35:09.880359 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" Dec 03 17:35:09 crc kubenswrapper[4787]: I1203 17:35:09.887297 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" Dec 03 17:35:09 crc kubenswrapper[4787]: I1203 17:35:09.961774 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scz6p\" (UniqueName: \"kubernetes.io/projected/c3b80bcb-cab7-4a0e-a761-ad5942435422-kube-api-access-scz6p\") pod \"c3b80bcb-cab7-4a0e-a761-ad5942435422\" (UID: \"c3b80bcb-cab7-4a0e-a761-ad5942435422\") " Dec 03 17:35:09 crc kubenswrapper[4787]: I1203 17:35:09.961885 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ec64d7f-0783-4048-a8ac-baa6de52ed78-operator-scripts\") pod \"4ec64d7f-0783-4048-a8ac-baa6de52ed78\" (UID: \"4ec64d7f-0783-4048-a8ac-baa6de52ed78\") " Dec 03 17:35:09 crc kubenswrapper[4787]: I1203 17:35:09.962047 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxvwl\" (UniqueName: \"kubernetes.io/projected/4ec64d7f-0783-4048-a8ac-baa6de52ed78-kube-api-access-gxvwl\") pod \"4ec64d7f-0783-4048-a8ac-baa6de52ed78\" (UID: \"4ec64d7f-0783-4048-a8ac-baa6de52ed78\") " Dec 03 17:35:09 crc kubenswrapper[4787]: I1203 17:35:09.962077 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3b80bcb-cab7-4a0e-a761-ad5942435422-operator-scripts\") pod \"c3b80bcb-cab7-4a0e-a761-ad5942435422\" (UID: \"c3b80bcb-cab7-4a0e-a761-ad5942435422\") " Dec 03 17:35:09 crc kubenswrapper[4787]: I1203 17:35:09.963226 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ec64d7f-0783-4048-a8ac-baa6de52ed78-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ec64d7f-0783-4048-a8ac-baa6de52ed78" (UID: "4ec64d7f-0783-4048-a8ac-baa6de52ed78"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:09 crc kubenswrapper[4787]: I1203 17:35:09.964782 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c3b80bcb-cab7-4a0e-a761-ad5942435422-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c3b80bcb-cab7-4a0e-a761-ad5942435422" (UID: "c3b80bcb-cab7-4a0e-a761-ad5942435422"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:09 crc kubenswrapper[4787]: I1203 17:35:09.973756 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ec64d7f-0783-4048-a8ac-baa6de52ed78-kube-api-access-gxvwl" (OuterVolumeSpecName: "kube-api-access-gxvwl") pod "4ec64d7f-0783-4048-a8ac-baa6de52ed78" (UID: "4ec64d7f-0783-4048-a8ac-baa6de52ed78"). InnerVolumeSpecName "kube-api-access-gxvwl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:09 crc kubenswrapper[4787]: I1203 17:35:09.986692 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c3b80bcb-cab7-4a0e-a761-ad5942435422-kube-api-access-scz6p" (OuterVolumeSpecName: "kube-api-access-scz6p") pod "c3b80bcb-cab7-4a0e-a761-ad5942435422" (UID: "c3b80bcb-cab7-4a0e-a761-ad5942435422"). InnerVolumeSpecName "kube-api-access-scz6p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:10 crc kubenswrapper[4787]: I1203 17:35:10.066493 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scz6p\" (UniqueName: \"kubernetes.io/projected/c3b80bcb-cab7-4a0e-a761-ad5942435422-kube-api-access-scz6p\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:10 crc kubenswrapper[4787]: I1203 17:35:10.066544 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ec64d7f-0783-4048-a8ac-baa6de52ed78-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:10 crc kubenswrapper[4787]: I1203 17:35:10.066558 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxvwl\" (UniqueName: \"kubernetes.io/projected/4ec64d7f-0783-4048-a8ac-baa6de52ed78-kube-api-access-gxvwl\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:10 crc kubenswrapper[4787]: I1203 17:35:10.066571 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c3b80bcb-cab7-4a0e-a761-ad5942435422-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:10 crc kubenswrapper[4787]: I1203 17:35:10.411080 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" event={"ID":"4ec64d7f-0783-4048-a8ac-baa6de52ed78","Type":"ContainerDied","Data":"a032b1313c52f1d5b2f63d146ea40e30f79eb8076a9cbf38f95f8ae0f92e2b94"} Dec 03 17:35:10 crc kubenswrapper[4787]: I1203 17:35:10.411159 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a032b1313c52f1d5b2f63d146ea40e30f79eb8076a9cbf38f95f8ae0f92e2b94" Dec 03 17:35:10 crc kubenswrapper[4787]: I1203 17:35:10.411091 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj" Dec 03 17:35:10 crc kubenswrapper[4787]: I1203 17:35:10.412728 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" event={"ID":"c3b80bcb-cab7-4a0e-a761-ad5942435422","Type":"ContainerDied","Data":"4116a2ec5d4e3cc7f4e9634604496484d1daed2044848dc5cf6a5b0ed52629f5"} Dec 03 17:35:10 crc kubenswrapper[4787]: I1203 17:35:10.412761 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4116a2ec5d4e3cc7f4e9634604496484d1daed2044848dc5cf6a5b0ed52629f5" Dec 03 17:35:10 crc kubenswrapper[4787]: I1203 17:35:10.412829 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-960e-account-create-update-8dsds" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.613875 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.616847 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-prdc5" podUID="2db4ba58-5d6c-42ff-b386-7765170d9113" containerName="ovn-controller" probeResult="failure" output=< Dec 03 17:35:12 crc kubenswrapper[4787]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 17:35:12 crc kubenswrapper[4787]: > Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.633037 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/106d2d96-ed76-4c59-9577-7086e56f9878-etc-swift\") pod \"swift-storage-0\" (UID: \"106d2d96-ed76-4c59-9577-7086e56f9878\") " pod="openstack/swift-storage-0" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.646941 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.697065 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-mc6r6" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.737043 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.908050 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-prdc5-config-nl5jd"] Dec 03 17:35:12 crc kubenswrapper[4787]: E1203 17:35:12.908639 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c3b80bcb-cab7-4a0e-a761-ad5942435422" containerName="mariadb-account-create-update" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.908661 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c3b80bcb-cab7-4a0e-a761-ad5942435422" containerName="mariadb-account-create-update" Dec 03 17:35:12 crc kubenswrapper[4787]: E1203 17:35:12.908677 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e4903b58-00fb-44e2-a2a1-60f89894b73c" containerName="console" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.908687 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e4903b58-00fb-44e2-a2a1-60f89894b73c" containerName="console" Dec 03 17:35:12 crc kubenswrapper[4787]: E1203 17:35:12.908704 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ec64d7f-0783-4048-a8ac-baa6de52ed78" containerName="mariadb-database-create" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.908713 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ec64d7f-0783-4048-a8ac-baa6de52ed78" containerName="mariadb-database-create" Dec 03 17:35:12 crc kubenswrapper[4787]: E1203 17:35:12.908762 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="973215cc-7718-421c-b374-49e64bb9af3e" containerName="swift-ring-rebalance" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.908772 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="973215cc-7718-421c-b374-49e64bb9af3e" containerName="swift-ring-rebalance" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.909037 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c3b80bcb-cab7-4a0e-a761-ad5942435422" containerName="mariadb-account-create-update" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.909056 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="973215cc-7718-421c-b374-49e64bb9af3e" containerName="swift-ring-rebalance" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.909075 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ec64d7f-0783-4048-a8ac-baa6de52ed78" containerName="mariadb-database-create" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.909088 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e4903b58-00fb-44e2-a2a1-60f89894b73c" containerName="console" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.909881 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.914366 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 17:35:12 crc kubenswrapper[4787]: I1203 17:35:12.923902 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-prdc5-config-nl5jd"] Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.023780 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.023850 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-log-ovn\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.024007 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run-ovn\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.024194 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-scripts\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.024275 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-additional-scripts\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.024313 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmbwp\" (UniqueName: \"kubernetes.io/projected/e7eac967-acd1-4b08-9d97-d901a877b896-kube-api-access-gmbwp\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.025133 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.126786 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run-ovn\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.126965 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-scripts\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.127009 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-additional-scripts\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.127050 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmbwp\" (UniqueName: \"kubernetes.io/projected/e7eac967-acd1-4b08-9d97-d901a877b896-kube-api-access-gmbwp\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.127208 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-log-ovn\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.127231 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.127548 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.128397 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-additional-scripts\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.129204 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-log-ovn\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.129529 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run-ovn\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.129584 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-scripts\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.159815 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmbwp\" (UniqueName: \"kubernetes.io/projected/e7eac967-acd1-4b08-9d97-d901a877b896-kube-api-access-gmbwp\") pod \"ovn-controller-prdc5-config-nl5jd\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:13 crc kubenswrapper[4787]: I1203 17:35:13.240764 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.818461 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.820790 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.835292 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.836771 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.865508 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " pod="openstack/mysqld-exporter-0" Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.865568 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-config-data\") pod \"mysqld-exporter-0\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " pod="openstack/mysqld-exporter-0" Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.865737 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrf2l\" (UniqueName: \"kubernetes.io/projected/561fe2dd-e048-4657-8c0c-2434fe09f35e-kube-api-access-xrf2l\") pod \"mysqld-exporter-0\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " pod="openstack/mysqld-exporter-0" Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.968261 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrf2l\" (UniqueName: \"kubernetes.io/projected/561fe2dd-e048-4657-8c0c-2434fe09f35e-kube-api-access-xrf2l\") pod \"mysqld-exporter-0\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " pod="openstack/mysqld-exporter-0" Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.968775 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " pod="openstack/mysqld-exporter-0" Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.968858 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-config-data\") pod \"mysqld-exporter-0\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " pod="openstack/mysqld-exporter-0" Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.977969 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " pod="openstack/mysqld-exporter-0" Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.979454 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-config-data\") pod \"mysqld-exporter-0\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " pod="openstack/mysqld-exporter-0" Dec 03 17:35:14 crc kubenswrapper[4787]: I1203 17:35:14.993955 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrf2l\" (UniqueName: \"kubernetes.io/projected/561fe2dd-e048-4657-8c0c-2434fe09f35e-kube-api-access-xrf2l\") pod \"mysqld-exporter-0\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " pod="openstack/mysqld-exporter-0" Dec 03 17:35:15 crc kubenswrapper[4787]: I1203 17:35:15.168254 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 17:35:15 crc kubenswrapper[4787]: I1203 17:35:15.460089 4787 generic.go:334] "Generic (PLEG): container finished" podID="d73396bd-2e1f-4c20-8327-610c5d783373" containerID="e140a21aafabe034cbd6828830bc11cbed6f27b23fdfcc9449ef640e44a0560b" exitCode=0 Dec 03 17:35:15 crc kubenswrapper[4787]: I1203 17:35:15.460157 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d73396bd-2e1f-4c20-8327-610c5d783373","Type":"ContainerDied","Data":"e140a21aafabe034cbd6828830bc11cbed6f27b23fdfcc9449ef640e44a0560b"} Dec 03 17:35:15 crc kubenswrapper[4787]: I1203 17:35:15.462379 4787 generic.go:334] "Generic (PLEG): container finished" podID="3180168d-9dad-441d-86bd-9d1801ef629f" containerID="2cad526e0647b21c366d8989df027fd6a0351ec3a648a935c487d2d379ae49af" exitCode=0 Dec 03 17:35:15 crc kubenswrapper[4787]: I1203 17:35:15.462410 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3180168d-9dad-441d-86bd-9d1801ef629f","Type":"ContainerDied","Data":"2cad526e0647b21c366d8989df027fd6a0351ec3a648a935c487d2d379ae49af"} Dec 03 17:35:17 crc kubenswrapper[4787]: I1203 17:35:17.625511 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-prdc5" podUID="2db4ba58-5d6c-42ff-b386-7765170d9113" containerName="ovn-controller" probeResult="failure" output=< Dec 03 17:35:17 crc kubenswrapper[4787]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 03 17:35:17 crc kubenswrapper[4787]: > Dec 03 17:35:17 crc kubenswrapper[4787]: I1203 17:35:17.927396 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 17:35:17 crc kubenswrapper[4787]: W1203 17:35:17.927908 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod561fe2dd_e048_4657_8c0c_2434fe09f35e.slice/crio-d138d90d677cba4874a5b2433644142946aa173ee140f87812c4e6174c477742 WatchSource:0}: Error finding container d138d90d677cba4874a5b2433644142946aa173ee140f87812c4e6174c477742: Status 404 returned error can't find the container with id d138d90d677cba4874a5b2433644142946aa173ee140f87812c4e6174c477742 Dec 03 17:35:17 crc kubenswrapper[4787]: I1203 17:35:17.932082 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:35:18 crc kubenswrapper[4787]: I1203 17:35:18.010988 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-prdc5-config-nl5jd"] Dec 03 17:35:18 crc kubenswrapper[4787]: I1203 17:35:18.105071 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Dec 03 17:35:18 crc kubenswrapper[4787]: W1203 17:35:18.110731 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod106d2d96_ed76_4c59_9577_7086e56f9878.slice/crio-a53cc4277d24593ab2ee9330125544fd036300323eab0e2365142088f83e488a WatchSource:0}: Error finding container a53cc4277d24593ab2ee9330125544fd036300323eab0e2365142088f83e488a: Status 404 returned error can't find the container with id a53cc4277d24593ab2ee9330125544fd036300323eab0e2365142088f83e488a Dec 03 17:35:18 crc kubenswrapper[4787]: I1203 17:35:18.493342 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3180168d-9dad-441d-86bd-9d1801ef629f","Type":"ContainerStarted","Data":"b4744a19f52cb6b0397c6428dd54a30ebe8f002f3d0e4df0a4d07e3e17f6fea0"} Dec 03 17:35:18 crc kubenswrapper[4787]: I1203 17:35:18.494579 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"a53cc4277d24593ab2ee9330125544fd036300323eab0e2365142088f83e488a"} Dec 03 17:35:18 crc kubenswrapper[4787]: I1203 17:35:18.495697 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-prdc5-config-nl5jd" event={"ID":"e7eac967-acd1-4b08-9d97-d901a877b896","Type":"ContainerStarted","Data":"6a1c6d9a7969a200080db6d81fd5b88de4b3c162d9efaf68dad4c43e6e4af4f1"} Dec 03 17:35:18 crc kubenswrapper[4787]: I1203 17:35:18.496700 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"561fe2dd-e048-4657-8c0c-2434fe09f35e","Type":"ContainerStarted","Data":"d138d90d677cba4874a5b2433644142946aa173ee140f87812c4e6174c477742"} Dec 03 17:35:19 crc kubenswrapper[4787]: I1203 17:35:19.510645 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d73396bd-2e1f-4c20-8327-610c5d783373","Type":"ContainerStarted","Data":"cb713dd212ea4276428258b4b48c6b045aa425d46ef7ca7d7e569df15f21de58"} Dec 03 17:35:19 crc kubenswrapper[4787]: I1203 17:35:19.511434 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 17:35:19 crc kubenswrapper[4787]: I1203 17:35:19.515590 4787 generic.go:334] "Generic (PLEG): container finished" podID="e7eac967-acd1-4b08-9d97-d901a877b896" containerID="18805ad3848e3536a98c60875d07221f74240d37ddcd3b7b0b20b92b27d3d959" exitCode=0 Dec 03 17:35:19 crc kubenswrapper[4787]: I1203 17:35:19.515790 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-prdc5-config-nl5jd" event={"ID":"e7eac967-acd1-4b08-9d97-d901a877b896","Type":"ContainerDied","Data":"18805ad3848e3536a98c60875d07221f74240d37ddcd3b7b0b20b92b27d3d959"} Dec 03 17:35:19 crc kubenswrapper[4787]: I1203 17:35:19.526477 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:35:19 crc kubenswrapper[4787]: I1203 17:35:19.547184 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=41.319697126 podStartE2EDuration="1m17.547166963s" podCreationTimestamp="2025-12-03 17:34:02 +0000 UTC" firstStartedPulling="2025-12-03 17:34:05.031325082 +0000 UTC m=+1261.848796041" lastFinishedPulling="2025-12-03 17:34:41.258794919 +0000 UTC m=+1298.076265878" observedRunningTime="2025-12-03 17:35:19.545724805 +0000 UTC m=+1336.363195784" watchObservedRunningTime="2025-12-03 17:35:19.547166963 +0000 UTC m=+1336.364637922" Dec 03 17:35:19 crc kubenswrapper[4787]: I1203 17:35:19.613004 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=41.112430932 podStartE2EDuration="1m17.612954351s" podCreationTimestamp="2025-12-03 17:34:02 +0000 UTC" firstStartedPulling="2025-12-03 17:34:04.848379944 +0000 UTC m=+1261.665850903" lastFinishedPulling="2025-12-03 17:34:41.348903363 +0000 UTC m=+1298.166374322" observedRunningTime="2025-12-03 17:35:19.596408041 +0000 UTC m=+1336.413879050" watchObservedRunningTime="2025-12-03 17:35:19.612954351 +0000 UTC m=+1336.430425330" Dec 03 17:35:20 crc kubenswrapper[4787]: I1203 17:35:20.529867 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wcnd6" event={"ID":"948f9158-3e6d-4a5d-a8d4-1c9480a85f90","Type":"ContainerStarted","Data":"77f034cd17d81b7795902c5d95791ed7f11491750435688747e326e98228c78b"} Dec 03 17:35:20 crc kubenswrapper[4787]: I1203 17:35:20.554432 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-wcnd6" podStartSLOduration=2.267248961 podStartE2EDuration="17.554416163s" podCreationTimestamp="2025-12-03 17:35:03 +0000 UTC" firstStartedPulling="2025-12-03 17:35:04.157056956 +0000 UTC m=+1320.974527915" lastFinishedPulling="2025-12-03 17:35:19.444224158 +0000 UTC m=+1336.261695117" observedRunningTime="2025-12-03 17:35:20.549733649 +0000 UTC m=+1337.367204618" watchObservedRunningTime="2025-12-03 17:35:20.554416163 +0000 UTC m=+1337.371887122" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.710359 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.817357 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-scripts\") pod \"e7eac967-acd1-4b08-9d97-d901a877b896\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.819324 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-scripts" (OuterVolumeSpecName: "scripts") pod "e7eac967-acd1-4b08-9d97-d901a877b896" (UID: "e7eac967-acd1-4b08-9d97-d901a877b896"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.819503 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-additional-scripts\") pod \"e7eac967-acd1-4b08-9d97-d901a877b896\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.822466 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "e7eac967-acd1-4b08-9d97-d901a877b896" (UID: "e7eac967-acd1-4b08-9d97-d901a877b896"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.822542 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmbwp\" (UniqueName: \"kubernetes.io/projected/e7eac967-acd1-4b08-9d97-d901a877b896-kube-api-access-gmbwp\") pod \"e7eac967-acd1-4b08-9d97-d901a877b896\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.823404 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run-ovn\") pod \"e7eac967-acd1-4b08-9d97-d901a877b896\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.823538 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "e7eac967-acd1-4b08-9d97-d901a877b896" (UID: "e7eac967-acd1-4b08-9d97-d901a877b896"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.825273 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run\") pod \"e7eac967-acd1-4b08-9d97-d901a877b896\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.825426 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run" (OuterVolumeSpecName: "var-run") pod "e7eac967-acd1-4b08-9d97-d901a877b896" (UID: "e7eac967-acd1-4b08-9d97-d901a877b896"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.825495 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-log-ovn\") pod \"e7eac967-acd1-4b08-9d97-d901a877b896\" (UID: \"e7eac967-acd1-4b08-9d97-d901a877b896\") " Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.825616 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "e7eac967-acd1-4b08-9d97-d901a877b896" (UID: "e7eac967-acd1-4b08-9d97-d901a877b896"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.826397 4787 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.826426 4787 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.826439 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.826452 4787 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/e7eac967-acd1-4b08-9d97-d901a877b896-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.826464 4787 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/e7eac967-acd1-4b08-9d97-d901a877b896-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.828995 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7eac967-acd1-4b08-9d97-d901a877b896-kube-api-access-gmbwp" (OuterVolumeSpecName: "kube-api-access-gmbwp") pod "e7eac967-acd1-4b08-9d97-d901a877b896" (UID: "e7eac967-acd1-4b08-9d97-d901a877b896"). InnerVolumeSpecName "kube-api-access-gmbwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:21 crc kubenswrapper[4787]: I1203 17:35:21.928324 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmbwp\" (UniqueName: \"kubernetes.io/projected/e7eac967-acd1-4b08-9d97-d901a877b896-kube-api-access-gmbwp\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.552999 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3b68d021-5db8-4a87-92f2-e827bbdd2255","Type":"ContainerStarted","Data":"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3"} Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.556219 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"561fe2dd-e048-4657-8c0c-2434fe09f35e","Type":"ContainerStarted","Data":"b4de70be6d3d549766d0ae5211c21fd50c58ada0b50b6cdf1f975662db978161"} Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.560339 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"7254dd6999611721864005576801cc4869b53606bc6fe359d31f19eadd69a23a"} Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.560394 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"227c1e9abb3966d0121a994483900ee3a1fcdab0f1d8a1ab5e4feaf3afe5db74"} Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.560408 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"a94eace9bbea7d277fa0f1b22b90991e20794d2e7a5095c378f2102c5a944f16"} Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.560420 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"60b2a237b3788a730d4cf517d4cc0fc7bac24378319a8296abd2be9ee5fc9750"} Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.562206 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-prdc5-config-nl5jd" event={"ID":"e7eac967-acd1-4b08-9d97-d901a877b896","Type":"ContainerDied","Data":"6a1c6d9a7969a200080db6d81fd5b88de4b3c162d9efaf68dad4c43e6e4af4f1"} Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.562255 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a1c6d9a7969a200080db6d81fd5b88de4b3c162d9efaf68dad4c43e6e4af4f1" Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.562266 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-prdc5-config-nl5jd" Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.587438 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=23.427678678 podStartE2EDuration="1m13.587415047s" podCreationTimestamp="2025-12-03 17:34:09 +0000 UTC" firstStartedPulling="2025-12-03 17:34:31.576599676 +0000 UTC m=+1288.394070635" lastFinishedPulling="2025-12-03 17:35:21.736336045 +0000 UTC m=+1338.553807004" observedRunningTime="2025-12-03 17:35:22.58227509 +0000 UTC m=+1339.399746039" watchObservedRunningTime="2025-12-03 17:35:22.587415047 +0000 UTC m=+1339.404886006" Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.602171 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=4.837024025 podStartE2EDuration="8.602148368s" podCreationTimestamp="2025-12-03 17:35:14 +0000 UTC" firstStartedPulling="2025-12-03 17:35:17.931760504 +0000 UTC m=+1334.749231463" lastFinishedPulling="2025-12-03 17:35:21.696884847 +0000 UTC m=+1338.514355806" observedRunningTime="2025-12-03 17:35:22.600435713 +0000 UTC m=+1339.417906672" watchObservedRunningTime="2025-12-03 17:35:22.602148368 +0000 UTC m=+1339.419619327" Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.670963 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-prdc5" Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.818209 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-prdc5-config-nl5jd"] Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.823659 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-prdc5-config-nl5jd"] Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.940920 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-prdc5-config-5lsjd"] Dec 03 17:35:22 crc kubenswrapper[4787]: E1203 17:35:22.941457 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e7eac967-acd1-4b08-9d97-d901a877b896" containerName="ovn-config" Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.941482 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e7eac967-acd1-4b08-9d97-d901a877b896" containerName="ovn-config" Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.941795 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e7eac967-acd1-4b08-9d97-d901a877b896" containerName="ovn-config" Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.942750 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.953058 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 03 17:35:22 crc kubenswrapper[4787]: I1203 17:35:22.974286 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-prdc5-config-5lsjd"] Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.055679 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-additional-scripts\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.055760 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run-ovn\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.055914 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-log-ovn\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.055942 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6brqv\" (UniqueName: \"kubernetes.io/projected/10661a83-d984-42e9-9f5c-d6086181aa0c-kube-api-access-6brqv\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.055971 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.056009 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-scripts\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.157430 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-log-ovn\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.157508 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6brqv\" (UniqueName: \"kubernetes.io/projected/10661a83-d984-42e9-9f5c-d6086181aa0c-kube-api-access-6brqv\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.157542 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.157573 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-scripts\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.157644 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-additional-scripts\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.157689 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run-ovn\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.157743 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-log-ovn\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.157818 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.157845 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run-ovn\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.158829 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-additional-scripts\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.160532 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-scripts\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.186776 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6brqv\" (UniqueName: \"kubernetes.io/projected/10661a83-d984-42e9-9f5c-d6086181aa0c-kube-api-access-6brqv\") pod \"ovn-controller-prdc5-config-5lsjd\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.305840 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.781208 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7eac967-acd1-4b08-9d97-d901a877b896" path="/var/lib/kubelet/pods/e7eac967-acd1-4b08-9d97-d901a877b896/volumes" Dec 03 17:35:23 crc kubenswrapper[4787]: I1203 17:35:23.784407 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-prdc5-config-5lsjd"] Dec 03 17:35:23 crc kubenswrapper[4787]: W1203 17:35:23.956492 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10661a83_d984_42e9_9f5c_d6086181aa0c.slice/crio-be4053a7d75deedc99abac8389034e88b00168e08f8f8f07d18df98af88c4391 WatchSource:0}: Error finding container be4053a7d75deedc99abac8389034e88b00168e08f8f8f07d18df98af88c4391: Status 404 returned error can't find the container with id be4053a7d75deedc99abac8389034e88b00168e08f8f8f07d18df98af88c4391 Dec 03 17:35:24 crc kubenswrapper[4787]: I1203 17:35:24.596248 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"1ad36ac7b50f122faeb8b52cce32812f19f132244d8c28f4766d511f32274809"} Dec 03 17:35:24 crc kubenswrapper[4787]: I1203 17:35:24.596958 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"7f9cac4a08ca235dfbd48b906d1a72993f706ec0fbdae67966a40c79d0edc915"} Dec 03 17:35:24 crc kubenswrapper[4787]: I1203 17:35:24.599779 4787 generic.go:334] "Generic (PLEG): container finished" podID="10661a83-d984-42e9-9f5c-d6086181aa0c" containerID="2e1ab55c40b297261a40fd7c4dfc3e647f8156cad557b8c44cb9f551fb6cfd6f" exitCode=0 Dec 03 17:35:24 crc kubenswrapper[4787]: I1203 17:35:24.599824 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-prdc5-config-5lsjd" event={"ID":"10661a83-d984-42e9-9f5c-d6086181aa0c","Type":"ContainerDied","Data":"2e1ab55c40b297261a40fd7c4dfc3e647f8156cad557b8c44cb9f551fb6cfd6f"} Dec 03 17:35:24 crc kubenswrapper[4787]: I1203 17:35:24.599851 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-prdc5-config-5lsjd" event={"ID":"10661a83-d984-42e9-9f5c-d6086181aa0c","Type":"ContainerStarted","Data":"be4053a7d75deedc99abac8389034e88b00168e08f8f8f07d18df98af88c4391"} Dec 03 17:35:25 crc kubenswrapper[4787]: I1203 17:35:25.618777 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"c106d2f1658701281cd866f785921a9f669e0a4e136e5c46f3c13f0f72d0559e"} Dec 03 17:35:25 crc kubenswrapper[4787]: I1203 17:35:25.876369 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:25 crc kubenswrapper[4787]: I1203 17:35:25.877500 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:25 crc kubenswrapper[4787]: I1203 17:35:25.885894 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.059375 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.113573 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run\") pod \"10661a83-d984-42e9-9f5c-d6086181aa0c\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.113837 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-scripts\") pod \"10661a83-d984-42e9-9f5c-d6086181aa0c\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.113915 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run-ovn\") pod \"10661a83-d984-42e9-9f5c-d6086181aa0c\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.113943 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-additional-scripts\") pod \"10661a83-d984-42e9-9f5c-d6086181aa0c\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.113980 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6brqv\" (UniqueName: \"kubernetes.io/projected/10661a83-d984-42e9-9f5c-d6086181aa0c-kube-api-access-6brqv\") pod \"10661a83-d984-42e9-9f5c-d6086181aa0c\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.114033 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-log-ovn\") pod \"10661a83-d984-42e9-9f5c-d6086181aa0c\" (UID: \"10661a83-d984-42e9-9f5c-d6086181aa0c\") " Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.114531 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "10661a83-d984-42e9-9f5c-d6086181aa0c" (UID: "10661a83-d984-42e9-9f5c-d6086181aa0c"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.114574 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run" (OuterVolumeSpecName: "var-run") pod "10661a83-d984-42e9-9f5c-d6086181aa0c" (UID: "10661a83-d984-42e9-9f5c-d6086181aa0c"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.115339 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "10661a83-d984-42e9-9f5c-d6086181aa0c" (UID: "10661a83-d984-42e9-9f5c-d6086181aa0c"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.118910 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "10661a83-d984-42e9-9f5c-d6086181aa0c" (UID: "10661a83-d984-42e9-9f5c-d6086181aa0c"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.119048 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-scripts" (OuterVolumeSpecName: "scripts") pod "10661a83-d984-42e9-9f5c-d6086181aa0c" (UID: "10661a83-d984-42e9-9f5c-d6086181aa0c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.149582 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10661a83-d984-42e9-9f5c-d6086181aa0c-kube-api-access-6brqv" (OuterVolumeSpecName: "kube-api-access-6brqv") pod "10661a83-d984-42e9-9f5c-d6086181aa0c" (UID: "10661a83-d984-42e9-9f5c-d6086181aa0c"). InnerVolumeSpecName "kube-api-access-6brqv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.216718 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.216770 4787 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.216786 4787 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/10661a83-d984-42e9-9f5c-d6086181aa0c-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.216798 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6brqv\" (UniqueName: \"kubernetes.io/projected/10661a83-d984-42e9-9f5c-d6086181aa0c-kube-api-access-6brqv\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.216809 4787 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.216818 4787 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/10661a83-d984-42e9-9f5c-d6086181aa0c-var-run\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.628238 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-prdc5-config-5lsjd" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.628260 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-prdc5-config-5lsjd" event={"ID":"10661a83-d984-42e9-9f5c-d6086181aa0c","Type":"ContainerDied","Data":"be4053a7d75deedc99abac8389034e88b00168e08f8f8f07d18df98af88c4391"} Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.628837 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be4053a7d75deedc99abac8389034e88b00168e08f8f8f07d18df98af88c4391" Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.633622 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"1ef4ed755d1f640fdf3d5cb4133783332b34015a206c8d91f4d29fc436eca186"} Dec 03 17:35:26 crc kubenswrapper[4787]: I1203 17:35:26.634301 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:27 crc kubenswrapper[4787]: I1203 17:35:27.144424 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-prdc5-config-5lsjd"] Dec 03 17:35:27 crc kubenswrapper[4787]: I1203 17:35:27.153558 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-prdc5-config-5lsjd"] Dec 03 17:35:27 crc kubenswrapper[4787]: I1203 17:35:27.696070 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"cf103db67469c408042dd5f3677062a979c1cf2215b44382b3013758b380b827"} Dec 03 17:35:27 crc kubenswrapper[4787]: I1203 17:35:27.784367 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="10661a83-d984-42e9-9f5c-d6086181aa0c" path="/var/lib/kubelet/pods/10661a83-d984-42e9-9f5c-d6086181aa0c/volumes" Dec 03 17:35:28 crc kubenswrapper[4787]: I1203 17:35:28.718274 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"a84d5fddfa685f75ba83fd3a053a9c55ddc4e2109629a7bc8bf5f6eac17e3678"} Dec 03 17:35:28 crc kubenswrapper[4787]: I1203 17:35:28.718702 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"e439d82e4a07cb3a140b49bcaf0218c20f149306410deb5ffd2b4df63c3a7810"} Dec 03 17:35:28 crc kubenswrapper[4787]: I1203 17:35:28.718714 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"4994f707aa6555d10e020987b150ec3ddb0ba0f736460811b5312775ffa6f68a"} Dec 03 17:35:28 crc kubenswrapper[4787]: I1203 17:35:28.718723 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"e582fcdc9931bc4ff5173b70dd7c16737ce7ac7ae69b824e9a80b4cfaf54ab69"} Dec 03 17:35:28 crc kubenswrapper[4787]: I1203 17:35:28.718733 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"2d3a3e28529a16c24a54c6f24770f9c5958a4462ea56e2e58a8991e7cc3a0fd8"} Dec 03 17:35:29 crc kubenswrapper[4787]: I1203 17:35:29.065604 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 17:35:29 crc kubenswrapper[4787]: I1203 17:35:29.735733 4787 generic.go:334] "Generic (PLEG): container finished" podID="948f9158-3e6d-4a5d-a8d4-1c9480a85f90" containerID="77f034cd17d81b7795902c5d95791ed7f11491750435688747e326e98228c78b" exitCode=0 Dec 03 17:35:29 crc kubenswrapper[4787]: I1203 17:35:29.735830 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wcnd6" event={"ID":"948f9158-3e6d-4a5d-a8d4-1c9480a85f90","Type":"ContainerDied","Data":"77f034cd17d81b7795902c5d95791ed7f11491750435688747e326e98228c78b"} Dec 03 17:35:29 crc kubenswrapper[4787]: I1203 17:35:29.745681 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"106d2d96-ed76-4c59-9577-7086e56f9878","Type":"ContainerStarted","Data":"0d91321893b3c9dcd3b6a240ebba58bd028a9e71545635ea7fb0be3410d0c2b3"} Dec 03 17:35:29 crc kubenswrapper[4787]: I1203 17:35:29.746081 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="thanos-sidecar" containerID="cri-o://d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3" gracePeriod=600 Dec 03 17:35:29 crc kubenswrapper[4787]: I1203 17:35:29.746057 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="prometheus" containerID="cri-o://782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb" gracePeriod=600 Dec 03 17:35:29 crc kubenswrapper[4787]: I1203 17:35:29.746086 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/prometheus-metric-storage-0" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="config-reloader" containerID="cri-o://19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213" gracePeriod=600 Dec 03 17:35:29 crc kubenswrapper[4787]: I1203 17:35:29.807342 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=41.791997036 podStartE2EDuration="50.807313737s" podCreationTimestamp="2025-12-03 17:34:39 +0000 UTC" firstStartedPulling="2025-12-03 17:35:18.11375457 +0000 UTC m=+1334.931225519" lastFinishedPulling="2025-12-03 17:35:27.129071261 +0000 UTC m=+1343.946542220" observedRunningTime="2025-12-03 17:35:29.796482459 +0000 UTC m=+1346.613953458" watchObservedRunningTime="2025-12-03 17:35:29.807313737 +0000 UTC m=+1346.624784726" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.061674 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rzsqx"] Dec 03 17:35:30 crc kubenswrapper[4787]: E1203 17:35:30.062075 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10661a83-d984-42e9-9f5c-d6086181aa0c" containerName="ovn-config" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.062089 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="10661a83-d984-42e9-9f5c-d6086181aa0c" containerName="ovn-config" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.066250 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="10661a83-d984-42e9-9f5c-d6086181aa0c" containerName="ovn-config" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.067382 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.076679 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rzsqx"] Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.077004 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.116324 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z8gp9\" (UniqueName: \"kubernetes.io/projected/7f93f1ab-a213-4c03-a0da-95f97e534f8b-kube-api-access-z8gp9\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.116462 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.116556 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-config\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.116634 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.116689 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.116737 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-svc\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.217937 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.218043 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.218081 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-svc\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.218107 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z8gp9\" (UniqueName: \"kubernetes.io/projected/7f93f1ab-a213-4c03-a0da-95f97e534f8b-kube-api-access-z8gp9\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.218146 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.218178 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-config\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.219149 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-config\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.219949 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-svc\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.220484 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-swift-storage-0\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.220594 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-sb\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.221704 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-nb\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.237874 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z8gp9\" (UniqueName: \"kubernetes.io/projected/7f93f1ab-a213-4c03-a0da-95f97e534f8b-kube-api-access-z8gp9\") pod \"dnsmasq-dns-764c5664d7-rzsqx\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.483327 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.666258 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.728759 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n62gz\" (UniqueName: \"kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-kube-api-access-n62gz\") pod \"3b68d021-5db8-4a87-92f2-e827bbdd2255\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.728874 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3b68d021-5db8-4a87-92f2-e827bbdd2255-config-out\") pod \"3b68d021-5db8-4a87-92f2-e827bbdd2255\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.728915 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-web-config\") pod \"3b68d021-5db8-4a87-92f2-e827bbdd2255\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.728934 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-config\") pod \"3b68d021-5db8-4a87-92f2-e827bbdd2255\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.728985 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-db\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"3b68d021-5db8-4a87-92f2-e827bbdd2255\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.729214 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-thanos-prometheus-http-client-file\") pod \"3b68d021-5db8-4a87-92f2-e827bbdd2255\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.729276 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3b68d021-5db8-4a87-92f2-e827bbdd2255-prometheus-metric-storage-rulefiles-0\") pod \"3b68d021-5db8-4a87-92f2-e827bbdd2255\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.729303 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-tls-assets\") pod \"3b68d021-5db8-4a87-92f2-e827bbdd2255\" (UID: \"3b68d021-5db8-4a87-92f2-e827bbdd2255\") " Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.732106 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3b68d021-5db8-4a87-92f2-e827bbdd2255-prometheus-metric-storage-rulefiles-0" (OuterVolumeSpecName: "prometheus-metric-storage-rulefiles-0") pod "3b68d021-5db8-4a87-92f2-e827bbdd2255" (UID: "3b68d021-5db8-4a87-92f2-e827bbdd2255"). InnerVolumeSpecName "prometheus-metric-storage-rulefiles-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.737386 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3b68d021-5db8-4a87-92f2-e827bbdd2255-config-out" (OuterVolumeSpecName: "config-out") pod "3b68d021-5db8-4a87-92f2-e827bbdd2255" (UID: "3b68d021-5db8-4a87-92f2-e827bbdd2255"). InnerVolumeSpecName "config-out". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.737423 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-kube-api-access-n62gz" (OuterVolumeSpecName: "kube-api-access-n62gz") pod "3b68d021-5db8-4a87-92f2-e827bbdd2255" (UID: "3b68d021-5db8-4a87-92f2-e827bbdd2255"). InnerVolumeSpecName "kube-api-access-n62gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.737503 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-tls-assets" (OuterVolumeSpecName: "tls-assets") pod "3b68d021-5db8-4a87-92f2-e827bbdd2255" (UID: "3b68d021-5db8-4a87-92f2-e827bbdd2255"). InnerVolumeSpecName "tls-assets". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.737530 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-thanos-prometheus-http-client-file" (OuterVolumeSpecName: "thanos-prometheus-http-client-file") pod "3b68d021-5db8-4a87-92f2-e827bbdd2255" (UID: "3b68d021-5db8-4a87-92f2-e827bbdd2255"). InnerVolumeSpecName "thanos-prometheus-http-client-file". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.738186 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-config" (OuterVolumeSpecName: "config") pod "3b68d021-5db8-4a87-92f2-e827bbdd2255" (UID: "3b68d021-5db8-4a87-92f2-e827bbdd2255"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.740521 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage08-crc" (OuterVolumeSpecName: "prometheus-metric-storage-db") pod "3b68d021-5db8-4a87-92f2-e827bbdd2255" (UID: "3b68d021-5db8-4a87-92f2-e827bbdd2255"). InnerVolumeSpecName "local-storage08-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.757623 4787 generic.go:334] "Generic (PLEG): container finished" podID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerID="d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3" exitCode=0 Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.757657 4787 generic.go:334] "Generic (PLEG): container finished" podID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerID="19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213" exitCode=0 Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.757674 4787 generic.go:334] "Generic (PLEG): container finished" podID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerID="782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb" exitCode=0 Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.758288 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3b68d021-5db8-4a87-92f2-e827bbdd2255","Type":"ContainerDied","Data":"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3"} Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.758310 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.758353 4787 scope.go:117] "RemoveContainer" containerID="d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.758342 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3b68d021-5db8-4a87-92f2-e827bbdd2255","Type":"ContainerDied","Data":"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213"} Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.758455 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3b68d021-5db8-4a87-92f2-e827bbdd2255","Type":"ContainerDied","Data":"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb"} Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.758468 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"3b68d021-5db8-4a87-92f2-e827bbdd2255","Type":"ContainerDied","Data":"e2aa77b4845468582d3647e9a42860dcb980d9e2227bdd8179a216ef2686ac89"} Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.768269 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-web-config" (OuterVolumeSpecName: "web-config") pod "3b68d021-5db8-4a87-92f2-e827bbdd2255" (UID: "3b68d021-5db8-4a87-92f2-e827bbdd2255"). InnerVolumeSpecName "web-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.782409 4787 scope.go:117] "RemoveContainer" containerID="19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.831472 4787 reconciler_common.go:293] "Volume detached for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/3b68d021-5db8-4a87-92f2-e827bbdd2255-config-out\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.831840 4787 reconciler_common.go:293] "Volume detached for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-web-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.831859 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.831886 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" " Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.831899 4787 reconciler_common.go:293] "Volume detached for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/3b68d021-5db8-4a87-92f2-e827bbdd2255-thanos-prometheus-http-client-file\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.831909 4787 reconciler_common.go:293] "Volume detached for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/3b68d021-5db8-4a87-92f2-e827bbdd2255-prometheus-metric-storage-rulefiles-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.831918 4787 reconciler_common.go:293] "Volume detached for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-tls-assets\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.831927 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n62gz\" (UniqueName: \"kubernetes.io/projected/3b68d021-5db8-4a87-92f2-e827bbdd2255-kube-api-access-n62gz\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.849414 4787 scope.go:117] "RemoveContainer" containerID="782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.856855 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage08-crc" (UniqueName: "kubernetes.io/local-volume/local-storage08-crc") on node "crc" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.932840 4787 scope.go:117] "RemoveContainer" containerID="eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.933489 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.955187 4787 scope.go:117] "RemoveContainer" containerID="d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3" Dec 03 17:35:30 crc kubenswrapper[4787]: E1203 17:35:30.956695 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3\": container with ID starting with d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3 not found: ID does not exist" containerID="d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.956737 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3"} err="failed to get container status \"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3\": rpc error: code = NotFound desc = could not find container \"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3\": container with ID starting with d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3 not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.956760 4787 scope.go:117] "RemoveContainer" containerID="19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213" Dec 03 17:35:30 crc kubenswrapper[4787]: E1203 17:35:30.957146 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213\": container with ID starting with 19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213 not found: ID does not exist" containerID="19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.957191 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213"} err="failed to get container status \"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213\": rpc error: code = NotFound desc = could not find container \"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213\": container with ID starting with 19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213 not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.957227 4787 scope.go:117] "RemoveContainer" containerID="782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb" Dec 03 17:35:30 crc kubenswrapper[4787]: E1203 17:35:30.957554 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb\": container with ID starting with 782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb not found: ID does not exist" containerID="782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.957576 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb"} err="failed to get container status \"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb\": rpc error: code = NotFound desc = could not find container \"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb\": container with ID starting with 782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.957593 4787 scope.go:117] "RemoveContainer" containerID="eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642" Dec 03 17:35:30 crc kubenswrapper[4787]: E1203 17:35:30.958196 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642\": container with ID starting with eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642 not found: ID does not exist" containerID="eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.958224 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642"} err="failed to get container status \"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642\": rpc error: code = NotFound desc = could not find container \"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642\": container with ID starting with eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642 not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.958244 4787 scope.go:117] "RemoveContainer" containerID="d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.965403 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3"} err="failed to get container status \"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3\": rpc error: code = NotFound desc = could not find container \"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3\": container with ID starting with d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3 not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.965436 4787 scope.go:117] "RemoveContainer" containerID="19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.970673 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213"} err="failed to get container status \"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213\": rpc error: code = NotFound desc = could not find container \"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213\": container with ID starting with 19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213 not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.970726 4787 scope.go:117] "RemoveContainer" containerID="782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.971145 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb"} err="failed to get container status \"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb\": rpc error: code = NotFound desc = could not find container \"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb\": container with ID starting with 782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.971186 4787 scope.go:117] "RemoveContainer" containerID="eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.971544 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642"} err="failed to get container status \"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642\": rpc error: code = NotFound desc = could not find container \"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642\": container with ID starting with eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642 not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.971612 4787 scope.go:117] "RemoveContainer" containerID="d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.971910 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3"} err="failed to get container status \"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3\": rpc error: code = NotFound desc = could not find container \"d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3\": container with ID starting with d2eebed1287ef458c6d38232e33377d3f61cc3623dc355930dd607ac84875ba3 not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.971944 4787 scope.go:117] "RemoveContainer" containerID="19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.972208 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213"} err="failed to get container status \"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213\": rpc error: code = NotFound desc = could not find container \"19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213\": container with ID starting with 19c3224d948c1ec2daeca5fc65b749956206b066070930d6d6b44ac6caffc213 not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.972234 4787 scope.go:117] "RemoveContainer" containerID="782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.972449 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb"} err="failed to get container status \"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb\": rpc error: code = NotFound desc = could not find container \"782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb\": container with ID starting with 782dd81f302cf33665ef6cb9ad43798fe4bbbeb85cd80facccb946de4c9c6abb not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.972474 4787 scope.go:117] "RemoveContainer" containerID="eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.973172 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642"} err="failed to get container status \"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642\": rpc error: code = NotFound desc = could not find container \"eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642\": container with ID starting with eb0c2ab6d28525353ffdc3c3b602e44c4717671ee3bf16290ea39987d89dc642 not found: ID does not exist" Dec 03 17:35:30 crc kubenswrapper[4787]: I1203 17:35:30.978759 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rzsqx"] Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.166607 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.189335 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.199473 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.227781 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 17:35:31 crc kubenswrapper[4787]: E1203 17:35:31.229048 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="init-config-reloader" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.229067 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="init-config-reloader" Dec 03 17:35:31 crc kubenswrapper[4787]: E1203 17:35:31.229078 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="prometheus" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.229084 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="prometheus" Dec 03 17:35:31 crc kubenswrapper[4787]: E1203 17:35:31.229097 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="config-reloader" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.229104 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="config-reloader" Dec 03 17:35:31 crc kubenswrapper[4787]: E1203 17:35:31.229113 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="948f9158-3e6d-4a5d-a8d4-1c9480a85f90" containerName="glance-db-sync" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.229118 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="948f9158-3e6d-4a5d-a8d4-1c9480a85f90" containerName="glance-db-sync" Dec 03 17:35:31 crc kubenswrapper[4787]: E1203 17:35:31.229145 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="thanos-sidecar" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.229151 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="thanos-sidecar" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.229322 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="config-reloader" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.229338 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="thanos-sidecar" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.229350 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="948f9158-3e6d-4a5d-a8d4-1c9480a85f90" containerName="glance-db-sync" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.229358 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" containerName="prometheus" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.230930 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.239586 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-db-sync-config-data\") pod \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.240073 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"prometheus-metric-storage-rulefiles-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.240322 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-thanos-prometheus-http-client-file" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.240451 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-web-config" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.240698 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.240936 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"metric-storage-prometheus-dockercfg-4g9vz" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.241200 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-metric-storage-prometheus-svc" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.242593 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-combined-ca-bundle\") pod \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.242729 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-config-data\") pod \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.242754 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqpwk\" (UniqueName: \"kubernetes.io/projected/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-kube-api-access-nqpwk\") pod \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\" (UID: \"948f9158-3e6d-4a5d-a8d4-1c9480a85f90\") " Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.246553 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "948f9158-3e6d-4a5d-a8d4-1c9480a85f90" (UID: "948f9158-3e6d-4a5d-a8d4-1c9480a85f90"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.251494 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-kube-api-access-nqpwk" (OuterVolumeSpecName: "kube-api-access-nqpwk") pod "948f9158-3e6d-4a5d-a8d4-1c9480a85f90" (UID: "948f9158-3e6d-4a5d-a8d4-1c9480a85f90"). InnerVolumeSpecName "kube-api-access-nqpwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.252465 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"prometheus-metric-storage-tls-assets-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.259455 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.288765 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "948f9158-3e6d-4a5d-a8d4-1c9480a85f90" (UID: "948f9158-3e6d-4a5d-a8d4-1c9480a85f90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.299649 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-config-data" (OuterVolumeSpecName: "config-data") pod "948f9158-3e6d-4a5d-a8d4-1c9480a85f90" (UID: "948f9158-3e6d-4a5d-a8d4-1c9480a85f90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.346458 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b7e89b75-951a-44b1-9362-fa52750d7450-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.346614 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.346675 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.346751 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.346833 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.346998 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b7e89b75-951a-44b1-9362-fa52750d7450-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.347104 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.347161 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5wzr8\" (UniqueName: \"kubernetes.io/projected/b7e89b75-951a-44b1-9362-fa52750d7450-kube-api-access-5wzr8\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.347222 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b7e89b75-951a-44b1-9362-fa52750d7450-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.347277 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.347337 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.347463 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.347488 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqpwk\" (UniqueName: \"kubernetes.io/projected/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-kube-api-access-nqpwk\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.347499 4787 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.347513 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/948f9158-3e6d-4a5d-a8d4-1c9480a85f90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.448602 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.448644 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.448676 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.448707 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.448748 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b7e89b75-951a-44b1-9362-fa52750d7450-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.448776 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.448801 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5wzr8\" (UniqueName: \"kubernetes.io/projected/b7e89b75-951a-44b1-9362-fa52750d7450-kube-api-access-5wzr8\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.448826 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b7e89b75-951a-44b1-9362-fa52750d7450-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.448846 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.448876 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.448932 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b7e89b75-951a-44b1-9362-fa52750d7450-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.450167 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.451220 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"prometheus-metric-storage-rulefiles-0\" (UniqueName: \"kubernetes.io/configmap/b7e89b75-951a-44b1-9362-fa52750d7450-prometheus-metric-storage-rulefiles-0\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.453997 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.454746 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"thanos-prometheus-http-client-file\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-thanos-prometheus-http-client-file\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.455310 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-out\" (UniqueName: \"kubernetes.io/empty-dir/b7e89b75-951a-44b1-9362-fa52750d7450-config-out\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.455328 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-assets\" (UniqueName: \"kubernetes.io/projected/b7e89b75-951a-44b1-9362-fa52750d7450-tls-assets\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.456711 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-web-config-tls-secret-key-cert-metric-storage-promethe-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.457197 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-web-config\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.458394 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-secret-combined-ca-bundle\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.460390 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\" (UniqueName: \"kubernetes.io/secret/b7e89b75-951a-44b1-9362-fa52750d7450-web-config-tls-secret-cert-cert-metric-storage-prometh-dc638c2d\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.467700 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5wzr8\" (UniqueName: \"kubernetes.io/projected/b7e89b75-951a-44b1-9362-fa52750d7450-kube-api-access-5wzr8\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.480996 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"prometheus-metric-storage-0\" (UID: \"b7e89b75-951a-44b1-9362-fa52750d7450\") " pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.571005 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.779433 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-wcnd6" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.785579 4787 generic.go:334] "Generic (PLEG): container finished" podID="7f93f1ab-a213-4c03-a0da-95f97e534f8b" containerID="9b3f983b2c020b4913c72273244795831c33fc22e92727e31bc3e4ff08c058cb" exitCode=0 Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.797786 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b68d021-5db8-4a87-92f2-e827bbdd2255" path="/var/lib/kubelet/pods/3b68d021-5db8-4a87-92f2-e827bbdd2255/volumes" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.802460 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-wcnd6" event={"ID":"948f9158-3e6d-4a5d-a8d4-1c9480a85f90","Type":"ContainerDied","Data":"ae5f0c500fe7b34fd2562552c411ca64313dc46aabeb746c19bb606ee9b520bc"} Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.802532 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ae5f0c500fe7b34fd2562552c411ca64313dc46aabeb746c19bb606ee9b520bc" Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.802547 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" event={"ID":"7f93f1ab-a213-4c03-a0da-95f97e534f8b","Type":"ContainerDied","Data":"9b3f983b2c020b4913c72273244795831c33fc22e92727e31bc3e4ff08c058cb"} Dec 03 17:35:31 crc kubenswrapper[4787]: I1203 17:35:31.802567 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" event={"ID":"7f93f1ab-a213-4c03-a0da-95f97e534f8b","Type":"ContainerStarted","Data":"ede7c8e01c6bb32b54a4d47734c2f5f4d4dd1ece554aadfda6a7c387e51595d5"} Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.068494 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/prometheus-metric-storage-0"] Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.238672 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rzsqx"] Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.305368 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-955rm"] Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.311391 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.349597 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-955rm"] Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.370152 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.370246 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.370365 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-config\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.370419 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzqvh\" (UniqueName: \"kubernetes.io/projected/ee7b5f10-8364-4019-990d-fd34a4c2eda5-kube-api-access-bzqvh\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.370436 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.370514 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.471685 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.471762 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.471801 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.471852 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-config\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.471869 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzqvh\" (UniqueName: \"kubernetes.io/projected/ee7b5f10-8364-4019-990d-fd34a4c2eda5-kube-api-access-bzqvh\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.471885 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.472926 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-swift-storage-0\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.473144 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-nb\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.473747 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-sb\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.473963 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-svc\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.474545 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-config\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.492262 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzqvh\" (UniqueName: \"kubernetes.io/projected/ee7b5f10-8364-4019-990d-fd34a4c2eda5-kube-api-access-bzqvh\") pod \"dnsmasq-dns-74f6bcbc87-955rm\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.716236 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.798146 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" event={"ID":"7f93f1ab-a213-4c03-a0da-95f97e534f8b","Type":"ContainerStarted","Data":"67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5"} Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.798306 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" podUID="7f93f1ab-a213-4c03-a0da-95f97e534f8b" containerName="dnsmasq-dns" containerID="cri-o://67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5" gracePeriod=10 Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.798355 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:32 crc kubenswrapper[4787]: I1203 17:35:32.800477 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7e89b75-951a-44b1-9362-fa52750d7450","Type":"ContainerStarted","Data":"f6193562e60befbb1c536bfca8c27d6de0c5df6cd2b5934d9cbe649ef28aafb5"} Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.191174 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" podStartSLOduration=3.191158009 podStartE2EDuration="3.191158009s" podCreationTimestamp="2025-12-03 17:35:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:35:32.824095427 +0000 UTC m=+1349.641566386" watchObservedRunningTime="2025-12-03 17:35:33.191158009 +0000 UTC m=+1350.008628968" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.199447 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-955rm"] Dec 03 17:35:33 crc kubenswrapper[4787]: W1203 17:35:33.203504 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podee7b5f10_8364_4019_990d_fd34a4c2eda5.slice/crio-fb73c9fa36acee53f9df14c114376742eec51f9f1a7a387739e65beff49f6264 WatchSource:0}: Error finding container fb73c9fa36acee53f9df14c114376742eec51f9f1a7a387739e65beff49f6264: Status 404 returned error can't find the container with id fb73c9fa36acee53f9df14c114376742eec51f9f1a7a387739e65beff49f6264 Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.347263 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.405696 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-sb\") pod \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.405800 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-nb\") pod \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.405838 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-config\") pod \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.405861 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z8gp9\" (UniqueName: \"kubernetes.io/projected/7f93f1ab-a213-4c03-a0da-95f97e534f8b-kube-api-access-z8gp9\") pod \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.405928 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-swift-storage-0\") pod \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.406005 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-svc\") pod \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\" (UID: \"7f93f1ab-a213-4c03-a0da-95f97e534f8b\") " Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.419251 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7f93f1ab-a213-4c03-a0da-95f97e534f8b-kube-api-access-z8gp9" (OuterVolumeSpecName: "kube-api-access-z8gp9") pod "7f93f1ab-a213-4c03-a0da-95f97e534f8b" (UID: "7f93f1ab-a213-4c03-a0da-95f97e534f8b"). InnerVolumeSpecName "kube-api-access-z8gp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.458947 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7f93f1ab-a213-4c03-a0da-95f97e534f8b" (UID: "7f93f1ab-a213-4c03-a0da-95f97e534f8b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.464200 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7f93f1ab-a213-4c03-a0da-95f97e534f8b" (UID: "7f93f1ab-a213-4c03-a0da-95f97e534f8b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.471214 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7f93f1ab-a213-4c03-a0da-95f97e534f8b" (UID: "7f93f1ab-a213-4c03-a0da-95f97e534f8b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.475486 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7f93f1ab-a213-4c03-a0da-95f97e534f8b" (UID: "7f93f1ab-a213-4c03-a0da-95f97e534f8b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.484179 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-config" (OuterVolumeSpecName: "config") pod "7f93f1ab-a213-4c03-a0da-95f97e534f8b" (UID: "7f93f1ab-a213-4c03-a0da-95f97e534f8b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.507739 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.507772 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.507782 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z8gp9\" (UniqueName: \"kubernetes.io/projected/7f93f1ab-a213-4c03-a0da-95f97e534f8b-kube-api-access-z8gp9\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.507792 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.507801 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.507808 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7f93f1ab-a213-4c03-a0da-95f97e534f8b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.816159 4787 generic.go:334] "Generic (PLEG): container finished" podID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerID="e24e83892e4407deb827e43d8dca0948a49be43613280b3fa68ad23b347e6a58" exitCode=0 Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.816235 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" event={"ID":"ee7b5f10-8364-4019-990d-fd34a4c2eda5","Type":"ContainerDied","Data":"e24e83892e4407deb827e43d8dca0948a49be43613280b3fa68ad23b347e6a58"} Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.816265 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" event={"ID":"ee7b5f10-8364-4019-990d-fd34a4c2eda5","Type":"ContainerStarted","Data":"fb73c9fa36acee53f9df14c114376742eec51f9f1a7a387739e65beff49f6264"} Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.819481 4787 generic.go:334] "Generic (PLEG): container finished" podID="7f93f1ab-a213-4c03-a0da-95f97e534f8b" containerID="67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5" exitCode=0 Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.819518 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" event={"ID":"7f93f1ab-a213-4c03-a0da-95f97e534f8b","Type":"ContainerDied","Data":"67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5"} Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.819543 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" event={"ID":"7f93f1ab-a213-4c03-a0da-95f97e534f8b","Type":"ContainerDied","Data":"ede7c8e01c6bb32b54a4d47734c2f5f4d4dd1ece554aadfda6a7c387e51595d5"} Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.819560 4787 scope.go:117] "RemoveContainer" containerID="67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.819568 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-764c5664d7-rzsqx" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.901109 4787 scope.go:117] "RemoveContainer" containerID="9b3f983b2c020b4913c72273244795831c33fc22e92727e31bc3e4ff08c058cb" Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.904178 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rzsqx"] Dec 03 17:35:33 crc kubenswrapper[4787]: I1203 17:35:33.913246 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-764c5664d7-rzsqx"] Dec 03 17:35:34 crc kubenswrapper[4787]: I1203 17:35:34.008451 4787 scope.go:117] "RemoveContainer" containerID="67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5" Dec 03 17:35:34 crc kubenswrapper[4787]: E1203 17:35:34.009034 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5\": container with ID starting with 67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5 not found: ID does not exist" containerID="67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5" Dec 03 17:35:34 crc kubenswrapper[4787]: I1203 17:35:34.009060 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5"} err="failed to get container status \"67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5\": rpc error: code = NotFound desc = could not find container \"67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5\": container with ID starting with 67c8e83abcc5cc53322bd15b4c60a80e7fbf5c4cdcc2634caa3f5e6bbb8174b5 not found: ID does not exist" Dec 03 17:35:34 crc kubenswrapper[4787]: I1203 17:35:34.009080 4787 scope.go:117] "RemoveContainer" containerID="9b3f983b2c020b4913c72273244795831c33fc22e92727e31bc3e4ff08c058cb" Dec 03 17:35:34 crc kubenswrapper[4787]: E1203 17:35:34.009258 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b3f983b2c020b4913c72273244795831c33fc22e92727e31bc3e4ff08c058cb\": container with ID starting with 9b3f983b2c020b4913c72273244795831c33fc22e92727e31bc3e4ff08c058cb not found: ID does not exist" containerID="9b3f983b2c020b4913c72273244795831c33fc22e92727e31bc3e4ff08c058cb" Dec 03 17:35:34 crc kubenswrapper[4787]: I1203 17:35:34.009274 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b3f983b2c020b4913c72273244795831c33fc22e92727e31bc3e4ff08c058cb"} err="failed to get container status \"9b3f983b2c020b4913c72273244795831c33fc22e92727e31bc3e4ff08c058cb\": rpc error: code = NotFound desc = could not find container \"9b3f983b2c020b4913c72273244795831c33fc22e92727e31bc3e4ff08c058cb\": container with ID starting with 9b3f983b2c020b4913c72273244795831c33fc22e92727e31bc3e4ff08c058cb not found: ID does not exist" Dec 03 17:35:34 crc kubenswrapper[4787]: I1203 17:35:34.046213 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:35:34 crc kubenswrapper[4787]: I1203 17:35:34.358183 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 17:35:34 crc kubenswrapper[4787]: I1203 17:35:34.830358 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" event={"ID":"ee7b5f10-8364-4019-990d-fd34a4c2eda5","Type":"ContainerStarted","Data":"4452da0c12d65077441132fee773247df42066fd552f22e4aec943fcbdc9a13a"} Dec 03 17:35:34 crc kubenswrapper[4787]: I1203 17:35:34.830798 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:34 crc kubenswrapper[4787]: I1203 17:35:34.869264 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" podStartSLOduration=2.869244413 podStartE2EDuration="2.869244413s" podCreationTimestamp="2025-12-03 17:35:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:35:34.862551475 +0000 UTC m=+1351.680022434" watchObservedRunningTime="2025-12-03 17:35:34.869244413 +0000 UTC m=+1351.686715372" Dec 03 17:35:35 crc kubenswrapper[4787]: I1203 17:35:35.781725 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7f93f1ab-a213-4c03-a0da-95f97e534f8b" path="/var/lib/kubelet/pods/7f93f1ab-a213-4c03-a0da-95f97e534f8b/volumes" Dec 03 17:35:35 crc kubenswrapper[4787]: I1203 17:35:35.844252 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7e89b75-951a-44b1-9362-fa52750d7450","Type":"ContainerStarted","Data":"e628463363d3335c455530efd456c735ea5d5443f10b1c00cf5ad291ed7928cc"} Dec 03 17:35:35 crc kubenswrapper[4787]: I1203 17:35:35.964741 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-create-9r7gn"] Dec 03 17:35:35 crc kubenswrapper[4787]: E1203 17:35:35.965259 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f93f1ab-a213-4c03-a0da-95f97e534f8b" containerName="dnsmasq-dns" Dec 03 17:35:35 crc kubenswrapper[4787]: I1203 17:35:35.965282 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f93f1ab-a213-4c03-a0da-95f97e534f8b" containerName="dnsmasq-dns" Dec 03 17:35:35 crc kubenswrapper[4787]: E1203 17:35:35.965315 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7f93f1ab-a213-4c03-a0da-95f97e534f8b" containerName="init" Dec 03 17:35:35 crc kubenswrapper[4787]: I1203 17:35:35.965323 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7f93f1ab-a213-4c03-a0da-95f97e534f8b" containerName="init" Dec 03 17:35:35 crc kubenswrapper[4787]: I1203 17:35:35.965510 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7f93f1ab-a213-4c03-a0da-95f97e534f8b" containerName="dnsmasq-dns" Dec 03 17:35:35 crc kubenswrapper[4787]: I1203 17:35:35.966257 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-9r7gn" Dec 03 17:35:35 crc kubenswrapper[4787]: I1203 17:35:35.974266 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-f622-account-create-update-mwk99"] Dec 03 17:35:35 crc kubenswrapper[4787]: I1203 17:35:35.975813 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f622-account-create-update-mwk99" Dec 03 17:35:35 crc kubenswrapper[4787]: I1203 17:35:35.979742 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 03 17:35:35 crc kubenswrapper[4787]: I1203 17:35:35.990319 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-9r7gn"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.042085 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f622-account-create-update-mwk99"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.073195 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j9b9c\" (UniqueName: \"kubernetes.io/projected/766cf6f2-1add-4b7e-a5d6-16facb6782dd-kube-api-access-j9b9c\") pod \"heat-db-create-9r7gn\" (UID: \"766cf6f2-1add-4b7e-a5d6-16facb6782dd\") " pod="openstack/heat-db-create-9r7gn" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.073302 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/766cf6f2-1add-4b7e-a5d6-16facb6782dd-operator-scripts\") pod \"heat-db-create-9r7gn\" (UID: \"766cf6f2-1add-4b7e-a5d6-16facb6782dd\") " pod="openstack/heat-db-create-9r7gn" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.083582 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-f4q94"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.085260 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f4q94" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.102700 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-f4q94"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.118959 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-d920-account-create-update-nqrk5"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.121281 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d920-account-create-update-nqrk5" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.124498 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-db-secret" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.131213 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-d920-account-create-update-nqrk5"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.179071 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlwqt\" (UniqueName: \"kubernetes.io/projected/b643325f-a8b6-4f76-ad35-52e0c4fd525e-kube-api-access-dlwqt\") pod \"barbican-f622-account-create-update-mwk99\" (UID: \"b643325f-a8b6-4f76-ad35-52e0c4fd525e\") " pod="openstack/barbican-f622-account-create-update-mwk99" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.179489 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/766cf6f2-1add-4b7e-a5d6-16facb6782dd-operator-scripts\") pod \"heat-db-create-9r7gn\" (UID: \"766cf6f2-1add-4b7e-a5d6-16facb6782dd\") " pod="openstack/heat-db-create-9r7gn" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.179523 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b643325f-a8b6-4f76-ad35-52e0c4fd525e-operator-scripts\") pod \"barbican-f622-account-create-update-mwk99\" (UID: \"b643325f-a8b6-4f76-ad35-52e0c4fd525e\") " pod="openstack/barbican-f622-account-create-update-mwk99" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.179698 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j9b9c\" (UniqueName: \"kubernetes.io/projected/766cf6f2-1add-4b7e-a5d6-16facb6782dd-kube-api-access-j9b9c\") pod \"heat-db-create-9r7gn\" (UID: \"766cf6f2-1add-4b7e-a5d6-16facb6782dd\") " pod="openstack/heat-db-create-9r7gn" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.181002 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/766cf6f2-1add-4b7e-a5d6-16facb6782dd-operator-scripts\") pod \"heat-db-create-9r7gn\" (UID: \"766cf6f2-1add-4b7e-a5d6-16facb6782dd\") " pod="openstack/heat-db-create-9r7gn" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.208284 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-4477n"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.209633 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4477n" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.219894 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-4477n"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.225943 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j9b9c\" (UniqueName: \"kubernetes.io/projected/766cf6f2-1add-4b7e-a5d6-16facb6782dd-kube-api-access-j9b9c\") pod \"heat-db-create-9r7gn\" (UID: \"766cf6f2-1add-4b7e-a5d6-16facb6782dd\") " pod="openstack/heat-db-create-9r7gn" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.262745 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-sx9j7"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.263992 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.267078 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.267445 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.267576 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-l49xh" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.267697 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.281327 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vqc8\" (UniqueName: \"kubernetes.io/projected/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-kube-api-access-7vqc8\") pod \"barbican-db-create-f4q94\" (UID: \"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679\") " pod="openstack/barbican-db-create-f4q94" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.281405 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-operator-scripts\") pod \"barbican-db-create-f4q94\" (UID: \"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679\") " pod="openstack/barbican-db-create-f4q94" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.281431 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0405ea9-d37c-40a5-98aa-86ed0d81c288-operator-scripts\") pod \"heat-d920-account-create-update-nqrk5\" (UID: \"c0405ea9-d37c-40a5-98aa-86ed0d81c288\") " pod="openstack/heat-d920-account-create-update-nqrk5" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.281464 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sgqz6\" (UniqueName: \"kubernetes.io/projected/c0405ea9-d37c-40a5-98aa-86ed0d81c288-kube-api-access-sgqz6\") pod \"heat-d920-account-create-update-nqrk5\" (UID: \"c0405ea9-d37c-40a5-98aa-86ed0d81c288\") " pod="openstack/heat-d920-account-create-update-nqrk5" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.281507 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlwqt\" (UniqueName: \"kubernetes.io/projected/b643325f-a8b6-4f76-ad35-52e0c4fd525e-kube-api-access-dlwqt\") pod \"barbican-f622-account-create-update-mwk99\" (UID: \"b643325f-a8b6-4f76-ad35-52e0c4fd525e\") " pod="openstack/barbican-f622-account-create-update-mwk99" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.281551 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b643325f-a8b6-4f76-ad35-52e0c4fd525e-operator-scripts\") pod \"barbican-f622-account-create-update-mwk99\" (UID: \"b643325f-a8b6-4f76-ad35-52e0c4fd525e\") " pod="openstack/barbican-f622-account-create-update-mwk99" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.282277 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b643325f-a8b6-4f76-ad35-52e0c4fd525e-operator-scripts\") pod \"barbican-f622-account-create-update-mwk99\" (UID: \"b643325f-a8b6-4f76-ad35-52e0c4fd525e\") " pod="openstack/barbican-f622-account-create-update-mwk99" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.283962 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-sx9j7"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.298631 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-9r7gn" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.322387 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlwqt\" (UniqueName: \"kubernetes.io/projected/b643325f-a8b6-4f76-ad35-52e0c4fd525e-kube-api-access-dlwqt\") pod \"barbican-f622-account-create-update-mwk99\" (UID: \"b643325f-a8b6-4f76-ad35-52e0c4fd525e\") " pod="openstack/barbican-f622-account-create-update-mwk99" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.382827 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-operator-scripts\") pod \"barbican-db-create-f4q94\" (UID: \"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679\") " pod="openstack/barbican-db-create-f4q94" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.382894 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0405ea9-d37c-40a5-98aa-86ed0d81c288-operator-scripts\") pod \"heat-d920-account-create-update-nqrk5\" (UID: \"c0405ea9-d37c-40a5-98aa-86ed0d81c288\") " pod="openstack/heat-d920-account-create-update-nqrk5" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.382943 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f36e27a-5a90-4f53-8b13-e42e134471f8-operator-scripts\") pod \"cinder-db-create-4477n\" (UID: \"3f36e27a-5a90-4f53-8b13-e42e134471f8\") " pod="openstack/cinder-db-create-4477n" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.382967 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sgqz6\" (UniqueName: \"kubernetes.io/projected/c0405ea9-d37c-40a5-98aa-86ed0d81c288-kube-api-access-sgqz6\") pod \"heat-d920-account-create-update-nqrk5\" (UID: \"c0405ea9-d37c-40a5-98aa-86ed0d81c288\") " pod="openstack/heat-d920-account-create-update-nqrk5" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.383156 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-config-data\") pod \"keystone-db-sync-sx9j7\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.383297 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8z454\" (UniqueName: \"kubernetes.io/projected/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-kube-api-access-8z454\") pod \"keystone-db-sync-sx9j7\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.383462 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vqc8\" (UniqueName: \"kubernetes.io/projected/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-kube-api-access-7vqc8\") pod \"barbican-db-create-f4q94\" (UID: \"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679\") " pod="openstack/barbican-db-create-f4q94" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.383495 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvbqd\" (UniqueName: \"kubernetes.io/projected/3f36e27a-5a90-4f53-8b13-e42e134471f8-kube-api-access-hvbqd\") pod \"cinder-db-create-4477n\" (UID: \"3f36e27a-5a90-4f53-8b13-e42e134471f8\") " pod="openstack/cinder-db-create-4477n" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.383544 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-combined-ca-bundle\") pod \"keystone-db-sync-sx9j7\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.383772 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-operator-scripts\") pod \"barbican-db-create-f4q94\" (UID: \"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679\") " pod="openstack/barbican-db-create-f4q94" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.384193 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0405ea9-d37c-40a5-98aa-86ed0d81c288-operator-scripts\") pod \"heat-d920-account-create-update-nqrk5\" (UID: \"c0405ea9-d37c-40a5-98aa-86ed0d81c288\") " pod="openstack/heat-d920-account-create-update-nqrk5" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.385483 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-8e2a-account-create-update-9z9kk"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.386732 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8e2a-account-create-update-9z9kk" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.390316 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.412076 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vqc8\" (UniqueName: \"kubernetes.io/projected/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-kube-api-access-7vqc8\") pod \"barbican-db-create-f4q94\" (UID: \"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679\") " pod="openstack/barbican-db-create-f4q94" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.412402 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8e2a-account-create-update-9z9kk"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.425388 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sgqz6\" (UniqueName: \"kubernetes.io/projected/c0405ea9-d37c-40a5-98aa-86ed0d81c288-kube-api-access-sgqz6\") pod \"heat-d920-account-create-update-nqrk5\" (UID: \"c0405ea9-d37c-40a5-98aa-86ed0d81c288\") " pod="openstack/heat-d920-account-create-update-nqrk5" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.446692 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d920-account-create-update-nqrk5" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.485323 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f36e27a-5a90-4f53-8b13-e42e134471f8-operator-scripts\") pod \"cinder-db-create-4477n\" (UID: \"3f36e27a-5a90-4f53-8b13-e42e134471f8\") " pod="openstack/cinder-db-create-4477n" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.485412 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-config-data\") pod \"keystone-db-sync-sx9j7\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.485451 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-operator-scripts\") pod \"cinder-8e2a-account-create-update-9z9kk\" (UID: \"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5\") " pod="openstack/cinder-8e2a-account-create-update-9z9kk" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.485480 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8z454\" (UniqueName: \"kubernetes.io/projected/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-kube-api-access-8z454\") pod \"keystone-db-sync-sx9j7\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.485518 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvbqd\" (UniqueName: \"kubernetes.io/projected/3f36e27a-5a90-4f53-8b13-e42e134471f8-kube-api-access-hvbqd\") pod \"cinder-db-create-4477n\" (UID: \"3f36e27a-5a90-4f53-8b13-e42e134471f8\") " pod="openstack/cinder-db-create-4477n" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.485541 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24qbz\" (UniqueName: \"kubernetes.io/projected/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-kube-api-access-24qbz\") pod \"cinder-8e2a-account-create-update-9z9kk\" (UID: \"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5\") " pod="openstack/cinder-8e2a-account-create-update-9z9kk" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.485697 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-combined-ca-bundle\") pod \"keystone-db-sync-sx9j7\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.486241 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f36e27a-5a90-4f53-8b13-e42e134471f8-operator-scripts\") pod \"cinder-db-create-4477n\" (UID: \"3f36e27a-5a90-4f53-8b13-e42e134471f8\") " pod="openstack/cinder-db-create-4477n" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.494040 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-combined-ca-bundle\") pod \"keystone-db-sync-sx9j7\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.494793 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-config-data\") pod \"keystone-db-sync-sx9j7\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.505801 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8z454\" (UniqueName: \"kubernetes.io/projected/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-kube-api-access-8z454\") pod \"keystone-db-sync-sx9j7\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.512143 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvbqd\" (UniqueName: \"kubernetes.io/projected/3f36e27a-5a90-4f53-8b13-e42e134471f8-kube-api-access-hvbqd\") pod \"cinder-db-create-4477n\" (UID: \"3f36e27a-5a90-4f53-8b13-e42e134471f8\") " pod="openstack/cinder-db-create-4477n" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.575477 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4477n" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.581862 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-zxhg9"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.585294 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zxhg9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.587863 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-operator-scripts\") pod \"cinder-8e2a-account-create-update-9z9kk\" (UID: \"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5\") " pod="openstack/cinder-8e2a-account-create-update-9z9kk" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.587928 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24qbz\" (UniqueName: \"kubernetes.io/projected/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-kube-api-access-24qbz\") pod \"cinder-8e2a-account-create-update-9z9kk\" (UID: \"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5\") " pod="openstack/cinder-8e2a-account-create-update-9z9kk" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.588748 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.588944 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-operator-scripts\") pod \"cinder-8e2a-account-create-update-9z9kk\" (UID: \"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5\") " pod="openstack/cinder-8e2a-account-create-update-9z9kk" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.622466 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f622-account-create-update-mwk99" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.629974 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zxhg9"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.640889 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24qbz\" (UniqueName: \"kubernetes.io/projected/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-kube-api-access-24qbz\") pod \"cinder-8e2a-account-create-update-9z9kk\" (UID: \"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5\") " pod="openstack/cinder-8e2a-account-create-update-9z9kk" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.667833 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-81ba-account-create-update-qfjj9"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.669226 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-81ba-account-create-update-qfjj9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.671381 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.691347 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjkdd\" (UniqueName: \"kubernetes.io/projected/4ec2c795-db12-41f1-a2cc-536471d01938-kube-api-access-fjkdd\") pod \"neutron-db-create-zxhg9\" (UID: \"4ec2c795-db12-41f1-a2cc-536471d01938\") " pod="openstack/neutron-db-create-zxhg9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.691431 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ec2c795-db12-41f1-a2cc-536471d01938-operator-scripts\") pod \"neutron-db-create-zxhg9\" (UID: \"4ec2c795-db12-41f1-a2cc-536471d01938\") " pod="openstack/neutron-db-create-zxhg9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.693361 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-81ba-account-create-update-qfjj9"] Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.708717 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f4q94" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.794523 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjkdd\" (UniqueName: \"kubernetes.io/projected/4ec2c795-db12-41f1-a2cc-536471d01938-kube-api-access-fjkdd\") pod \"neutron-db-create-zxhg9\" (UID: \"4ec2c795-db12-41f1-a2cc-536471d01938\") " pod="openstack/neutron-db-create-zxhg9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.794871 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ec2c795-db12-41f1-a2cc-536471d01938-operator-scripts\") pod \"neutron-db-create-zxhg9\" (UID: \"4ec2c795-db12-41f1-a2cc-536471d01938\") " pod="openstack/neutron-db-create-zxhg9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.794928 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82zck\" (UniqueName: \"kubernetes.io/projected/8ee1f7fe-90fd-4289-b280-f48d0076ce09-kube-api-access-82zck\") pod \"neutron-81ba-account-create-update-qfjj9\" (UID: \"8ee1f7fe-90fd-4289-b280-f48d0076ce09\") " pod="openstack/neutron-81ba-account-create-update-qfjj9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.794984 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ee1f7fe-90fd-4289-b280-f48d0076ce09-operator-scripts\") pod \"neutron-81ba-account-create-update-qfjj9\" (UID: \"8ee1f7fe-90fd-4289-b280-f48d0076ce09\") " pod="openstack/neutron-81ba-account-create-update-qfjj9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.795945 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ec2c795-db12-41f1-a2cc-536471d01938-operator-scripts\") pod \"neutron-db-create-zxhg9\" (UID: \"4ec2c795-db12-41f1-a2cc-536471d01938\") " pod="openstack/neutron-db-create-zxhg9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.797077 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8e2a-account-create-update-9z9kk" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.829688 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjkdd\" (UniqueName: \"kubernetes.io/projected/4ec2c795-db12-41f1-a2cc-536471d01938-kube-api-access-fjkdd\") pod \"neutron-db-create-zxhg9\" (UID: \"4ec2c795-db12-41f1-a2cc-536471d01938\") " pod="openstack/neutron-db-create-zxhg9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.896485 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82zck\" (UniqueName: \"kubernetes.io/projected/8ee1f7fe-90fd-4289-b280-f48d0076ce09-kube-api-access-82zck\") pod \"neutron-81ba-account-create-update-qfjj9\" (UID: \"8ee1f7fe-90fd-4289-b280-f48d0076ce09\") " pod="openstack/neutron-81ba-account-create-update-qfjj9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.896566 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ee1f7fe-90fd-4289-b280-f48d0076ce09-operator-scripts\") pod \"neutron-81ba-account-create-update-qfjj9\" (UID: \"8ee1f7fe-90fd-4289-b280-f48d0076ce09\") " pod="openstack/neutron-81ba-account-create-update-qfjj9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.898582 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ee1f7fe-90fd-4289-b280-f48d0076ce09-operator-scripts\") pod \"neutron-81ba-account-create-update-qfjj9\" (UID: \"8ee1f7fe-90fd-4289-b280-f48d0076ce09\") " pod="openstack/neutron-81ba-account-create-update-qfjj9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.922842 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zxhg9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.922963 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82zck\" (UniqueName: \"kubernetes.io/projected/8ee1f7fe-90fd-4289-b280-f48d0076ce09-kube-api-access-82zck\") pod \"neutron-81ba-account-create-update-qfjj9\" (UID: \"8ee1f7fe-90fd-4289-b280-f48d0076ce09\") " pod="openstack/neutron-81ba-account-create-update-qfjj9" Dec 03 17:35:36 crc kubenswrapper[4787]: I1203 17:35:36.972211 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-create-9r7gn"] Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.001523 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-81ba-account-create-update-qfjj9" Dec 03 17:35:37 crc kubenswrapper[4787]: W1203 17:35:37.019640 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod766cf6f2_1add_4b7e_a5d6_16facb6782dd.slice/crio-41cb03e50f0adaf6091c0d968df4576d0a116cf554b61bb49105f49db6bde2f9 WatchSource:0}: Error finding container 41cb03e50f0adaf6091c0d968df4576d0a116cf554b61bb49105f49db6bde2f9: Status 404 returned error can't find the container with id 41cb03e50f0adaf6091c0d968df4576d0a116cf554b61bb49105f49db6bde2f9 Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.187242 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-d920-account-create-update-nqrk5"] Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.479774 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-4477n"] Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.534464 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-f622-account-create-update-mwk99"] Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.579715 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-sx9j7"] Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.760531 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-f4q94"] Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.805101 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-8e2a-account-create-update-9z9kk"] Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.818590 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-zxhg9"] Dec 03 17:35:37 crc kubenswrapper[4787]: W1203 17:35:37.885809 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ec2c795_db12_41f1_a2cc_536471d01938.slice/crio-f361b494f9970cf9f47c8ddafdac283b91128c08757ef61caaf1dc8c532d0b77 WatchSource:0}: Error finding container f361b494f9970cf9f47c8ddafdac283b91128c08757ef61caaf1dc8c532d0b77: Status 404 returned error can't find the container with id f361b494f9970cf9f47c8ddafdac283b91128c08757ef61caaf1dc8c532d0b77 Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.891702 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-81ba-account-create-update-qfjj9"] Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.894236 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8e2a-account-create-update-9z9kk" event={"ID":"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5","Type":"ContainerStarted","Data":"0acbc057e853a4cbfd07de9759df8147d6ecb8a103a90d76479f46b54ab06fb7"} Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.896312 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-d920-account-create-update-nqrk5" event={"ID":"c0405ea9-d37c-40a5-98aa-86ed0d81c288","Type":"ContainerStarted","Data":"7f75a87be2eb969350afd79628c6a52fc83c0a2ce5c41dc3a78e36e7a4f23ca3"} Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.901394 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-sx9j7" event={"ID":"dadfdc4d-30a6-4edd-88a4-8835c6e433c0","Type":"ContainerStarted","Data":"3803d93b643cfb4b576bfc92a9ac70d95a28aeebfbc94c87800bb95af390939e"} Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.903122 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f622-account-create-update-mwk99" event={"ID":"b643325f-a8b6-4f76-ad35-52e0c4fd525e","Type":"ContainerStarted","Data":"437b701d7ade9ee16749d763f478d43ea09a6baa8b8c83dbd7acef3824b1e6c6"} Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.906100 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4477n" event={"ID":"3f36e27a-5a90-4f53-8b13-e42e134471f8","Type":"ContainerStarted","Data":"483fe6637b269d25ddd59d4abed72b031021db7fc842a677746c261c07a8f3d6"} Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.910058 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-9r7gn" event={"ID":"766cf6f2-1add-4b7e-a5d6-16facb6782dd","Type":"ContainerStarted","Data":"bbfbd7081f97a5b2201c799c4903d708f81a0e9cfc311ecd112967c9c896d002"} Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.910110 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-9r7gn" event={"ID":"766cf6f2-1add-4b7e-a5d6-16facb6782dd","Type":"ContainerStarted","Data":"41cb03e50f0adaf6091c0d968df4576d0a116cf554b61bb49105f49db6bde2f9"} Dec 03 17:35:37 crc kubenswrapper[4787]: I1203 17:35:37.915917 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-f4q94" event={"ID":"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679","Type":"ContainerStarted","Data":"9906a94742d3322e0dd5dea2b0e121d996ebb53db932911368dc94cc6977157c"} Dec 03 17:35:37 crc kubenswrapper[4787]: W1203 17:35:37.970389 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ee1f7fe_90fd_4289_b280_f48d0076ce09.slice/crio-e6c44a8797519bcebe5babeb7b0ec7b00111130acc36317f340873c0ad802e8a WatchSource:0}: Error finding container e6c44a8797519bcebe5babeb7b0ec7b00111130acc36317f340873c0ad802e8a: Status 404 returned error can't find the container with id e6c44a8797519bcebe5babeb7b0ec7b00111130acc36317f340873c0ad802e8a Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.933562 4787 generic.go:334] "Generic (PLEG): container finished" podID="4ec2c795-db12-41f1-a2cc-536471d01938" containerID="6f2e1cb7a6767cf507693445b32f7919569248618d8d0ea7829b4ac3c85d02c9" exitCode=0 Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.933674 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zxhg9" event={"ID":"4ec2c795-db12-41f1-a2cc-536471d01938","Type":"ContainerDied","Data":"6f2e1cb7a6767cf507693445b32f7919569248618d8d0ea7829b4ac3c85d02c9"} Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.934184 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zxhg9" event={"ID":"4ec2c795-db12-41f1-a2cc-536471d01938","Type":"ContainerStarted","Data":"f361b494f9970cf9f47c8ddafdac283b91128c08757ef61caaf1dc8c532d0b77"} Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.938207 4787 generic.go:334] "Generic (PLEG): container finished" podID="c0405ea9-d37c-40a5-98aa-86ed0d81c288" containerID="5d26d36c88fe30d95249b94ec41893577fc809ea40eef42fce8ab575360bf0dc" exitCode=0 Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.938259 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-d920-account-create-update-nqrk5" event={"ID":"c0405ea9-d37c-40a5-98aa-86ed0d81c288","Type":"ContainerDied","Data":"5d26d36c88fe30d95249b94ec41893577fc809ea40eef42fce8ab575360bf0dc"} Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.943327 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-81ba-account-create-update-qfjj9" event={"ID":"8ee1f7fe-90fd-4289-b280-f48d0076ce09","Type":"ContainerStarted","Data":"a8b56fb0ac3b3b1adea50394e76383617e9788f6ca5f9a25b53d165e1a15a749"} Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.943375 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-81ba-account-create-update-qfjj9" event={"ID":"8ee1f7fe-90fd-4289-b280-f48d0076ce09","Type":"ContainerStarted","Data":"e6c44a8797519bcebe5babeb7b0ec7b00111130acc36317f340873c0ad802e8a"} Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.951558 4787 generic.go:334] "Generic (PLEG): container finished" podID="b643325f-a8b6-4f76-ad35-52e0c4fd525e" containerID="42f5d5e3986f90c800097a400e9327f1fdd2e107d2cc8fab37c993e8910fbf8d" exitCode=0 Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.951657 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f622-account-create-update-mwk99" event={"ID":"b643325f-a8b6-4f76-ad35-52e0c4fd525e","Type":"ContainerDied","Data":"42f5d5e3986f90c800097a400e9327f1fdd2e107d2cc8fab37c993e8910fbf8d"} Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.956299 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-create-9r7gn" podStartSLOduration=3.956281208 podStartE2EDuration="3.956281208s" podCreationTimestamp="2025-12-03 17:35:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:35:37.935334674 +0000 UTC m=+1354.752805643" watchObservedRunningTime="2025-12-03 17:35:38.956281208 +0000 UTC m=+1355.773752167" Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.957311 4787 generic.go:334] "Generic (PLEG): container finished" podID="3f36e27a-5a90-4f53-8b13-e42e134471f8" containerID="47331be765582c7502e1bb5e32f953f0bbab28d27d64505f9058a7a31f75862d" exitCode=0 Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.957356 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4477n" event={"ID":"3f36e27a-5a90-4f53-8b13-e42e134471f8","Type":"ContainerDied","Data":"47331be765582c7502e1bb5e32f953f0bbab28d27d64505f9058a7a31f75862d"} Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.958812 4787 generic.go:334] "Generic (PLEG): container finished" podID="766cf6f2-1add-4b7e-a5d6-16facb6782dd" containerID="bbfbd7081f97a5b2201c799c4903d708f81a0e9cfc311ecd112967c9c896d002" exitCode=0 Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.958872 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-9r7gn" event={"ID":"766cf6f2-1add-4b7e-a5d6-16facb6782dd","Type":"ContainerDied","Data":"bbfbd7081f97a5b2201c799c4903d708f81a0e9cfc311ecd112967c9c896d002"} Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.961480 4787 generic.go:334] "Generic (PLEG): container finished" podID="f43e786a-e9ff-4ca7-a0d6-f9b9ce928679" containerID="a1f0db31f1451a1e3dc6428e9a1678b70a9ab04ce4dea0d76978ed185d13a824" exitCode=0 Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.961633 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-f4q94" event={"ID":"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679","Type":"ContainerDied","Data":"a1f0db31f1451a1e3dc6428e9a1678b70a9ab04ce4dea0d76978ed185d13a824"} Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.966539 4787 generic.go:334] "Generic (PLEG): container finished" podID="d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5" containerID="ac282b9d9d2a8599bfff0d29214bcdaa84ddda0651f2a9d8b262a9922a498891" exitCode=0 Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.966591 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8e2a-account-create-update-9z9kk" event={"ID":"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5","Type":"ContainerDied","Data":"ac282b9d9d2a8599bfff0d29214bcdaa84ddda0651f2a9d8b262a9922a498891"} Dec 03 17:35:38 crc kubenswrapper[4787]: I1203 17:35:38.970998 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-81ba-account-create-update-qfjj9" podStartSLOduration=2.970986029 podStartE2EDuration="2.970986029s" podCreationTimestamp="2025-12-03 17:35:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:35:38.961113197 +0000 UTC m=+1355.778584166" watchObservedRunningTime="2025-12-03 17:35:38.970986029 +0000 UTC m=+1355.788456988" Dec 03 17:35:39 crc kubenswrapper[4787]: I1203 17:35:39.980443 4787 generic.go:334] "Generic (PLEG): container finished" podID="8ee1f7fe-90fd-4289-b280-f48d0076ce09" containerID="a8b56fb0ac3b3b1adea50394e76383617e9788f6ca5f9a25b53d165e1a15a749" exitCode=0 Dec 03 17:35:39 crc kubenswrapper[4787]: I1203 17:35:39.980559 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-81ba-account-create-update-qfjj9" event={"ID":"8ee1f7fe-90fd-4289-b280-f48d0076ce09","Type":"ContainerDied","Data":"a8b56fb0ac3b3b1adea50394e76383617e9788f6ca5f9a25b53d165e1a15a749"} Dec 03 17:35:42 crc kubenswrapper[4787]: I1203 17:35:42.003350 4787 generic.go:334] "Generic (PLEG): container finished" podID="b7e89b75-951a-44b1-9362-fa52750d7450" containerID="e628463363d3335c455530efd456c735ea5d5443f10b1c00cf5ad291ed7928cc" exitCode=0 Dec 03 17:35:42 crc kubenswrapper[4787]: I1203 17:35:42.003625 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7e89b75-951a-44b1-9362-fa52750d7450","Type":"ContainerDied","Data":"e628463363d3335c455530efd456c735ea5d5443f10b1c00cf5ad291ed7928cc"} Dec 03 17:35:42 crc kubenswrapper[4787]: I1203 17:35:42.718229 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:35:42 crc kubenswrapper[4787]: I1203 17:35:42.774836 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kspp6"] Dec 03 17:35:42 crc kubenswrapper[4787]: I1203 17:35:42.775130 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-698758b865-kspp6" podUID="d0bc011b-984e-4e56-a3c2-b5ea04a74a73" containerName="dnsmasq-dns" containerID="cri-o://783be89bd64ad820febf63d687389dcb5e50ed3b1d5295a09290c8de26b99293" gracePeriod=10 Dec 03 17:35:42 crc kubenswrapper[4787]: I1203 17:35:42.972613 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-9r7gn" Dec 03 17:35:42 crc kubenswrapper[4787]: I1203 17:35:42.997663 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4477n" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.027710 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zxhg9" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.036865 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f4q94" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.044723 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-8e2a-account-create-update-9z9kk" event={"ID":"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5","Type":"ContainerDied","Data":"0acbc057e853a4cbfd07de9759df8147d6ecb8a103a90d76479f46b54ab06fb7"} Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.044762 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0acbc057e853a4cbfd07de9759df8147d6ecb8a103a90d76479f46b54ab06fb7" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.045479 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f622-account-create-update-mwk99" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.046895 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-zxhg9" event={"ID":"4ec2c795-db12-41f1-a2cc-536471d01938","Type":"ContainerDied","Data":"f361b494f9970cf9f47c8ddafdac283b91128c08757ef61caaf1dc8c532d0b77"} Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.046921 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f361b494f9970cf9f47c8ddafdac283b91128c08757ef61caaf1dc8c532d0b77" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.046940 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-zxhg9" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.048802 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-f622-account-create-update-mwk99" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.048791 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-f622-account-create-update-mwk99" event={"ID":"b643325f-a8b6-4f76-ad35-52e0c4fd525e","Type":"ContainerDied","Data":"437b701d7ade9ee16749d763f478d43ea09a6baa8b8c83dbd7acef3824b1e6c6"} Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.048850 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="437b701d7ade9ee16749d763f478d43ea09a6baa8b8c83dbd7acef3824b1e6c6" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.050864 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-4477n" event={"ID":"3f36e27a-5a90-4f53-8b13-e42e134471f8","Type":"ContainerDied","Data":"483fe6637b269d25ddd59d4abed72b031021db7fc842a677746c261c07a8f3d6"} Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.050889 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="483fe6637b269d25ddd59d4abed72b031021db7fc842a677746c261c07a8f3d6" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.050912 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-4477n" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.051256 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8e2a-account-create-update-9z9kk" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.052075 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-create-9r7gn" event={"ID":"766cf6f2-1add-4b7e-a5d6-16facb6782dd","Type":"ContainerDied","Data":"41cb03e50f0adaf6091c0d968df4576d0a116cf554b61bb49105f49db6bde2f9"} Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.052091 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="41cb03e50f0adaf6091c0d968df4576d0a116cf554b61bb49105f49db6bde2f9" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.052144 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-create-9r7gn" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.053723 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-f4q94" event={"ID":"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679","Type":"ContainerDied","Data":"9906a94742d3322e0dd5dea2b0e121d996ebb53db932911368dc94cc6977157c"} Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.053752 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9906a94742d3322e0dd5dea2b0e121d996ebb53db932911368dc94cc6977157c" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.053795 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f4q94" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.056283 4787 generic.go:334] "Generic (PLEG): container finished" podID="d0bc011b-984e-4e56-a3c2-b5ea04a74a73" containerID="783be89bd64ad820febf63d687389dcb5e50ed3b1d5295a09290c8de26b99293" exitCode=0 Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.056351 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kspp6" event={"ID":"d0bc011b-984e-4e56-a3c2-b5ea04a74a73","Type":"ContainerDied","Data":"783be89bd64ad820febf63d687389dcb5e50ed3b1d5295a09290c8de26b99293"} Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.064719 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-d920-account-create-update-nqrk5" event={"ID":"c0405ea9-d37c-40a5-98aa-86ed0d81c288","Type":"ContainerDied","Data":"7f75a87be2eb969350afd79628c6a52fc83c0a2ce5c41dc3a78e36e7a4f23ca3"} Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.064747 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f75a87be2eb969350afd79628c6a52fc83c0a2ce5c41dc3a78e36e7a4f23ca3" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.065808 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/766cf6f2-1add-4b7e-a5d6-16facb6782dd-operator-scripts\") pod \"766cf6f2-1add-4b7e-a5d6-16facb6782dd\" (UID: \"766cf6f2-1add-4b7e-a5d6-16facb6782dd\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.065850 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvbqd\" (UniqueName: \"kubernetes.io/projected/3f36e27a-5a90-4f53-8b13-e42e134471f8-kube-api-access-hvbqd\") pod \"3f36e27a-5a90-4f53-8b13-e42e134471f8\" (UID: \"3f36e27a-5a90-4f53-8b13-e42e134471f8\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.065935 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j9b9c\" (UniqueName: \"kubernetes.io/projected/766cf6f2-1add-4b7e-a5d6-16facb6782dd-kube-api-access-j9b9c\") pod \"766cf6f2-1add-4b7e-a5d6-16facb6782dd\" (UID: \"766cf6f2-1add-4b7e-a5d6-16facb6782dd\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.065971 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f36e27a-5a90-4f53-8b13-e42e134471f8-operator-scripts\") pod \"3f36e27a-5a90-4f53-8b13-e42e134471f8\" (UID: \"3f36e27a-5a90-4f53-8b13-e42e134471f8\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.066424 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/766cf6f2-1add-4b7e-a5d6-16facb6782dd-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "766cf6f2-1add-4b7e-a5d6-16facb6782dd" (UID: "766cf6f2-1add-4b7e-a5d6-16facb6782dd"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.066827 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f36e27a-5a90-4f53-8b13-e42e134471f8-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3f36e27a-5a90-4f53-8b13-e42e134471f8" (UID: "3f36e27a-5a90-4f53-8b13-e42e134471f8"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.067347 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3f36e27a-5a90-4f53-8b13-e42e134471f8-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.067368 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/766cf6f2-1add-4b7e-a5d6-16facb6782dd-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.071123 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d920-account-create-update-nqrk5" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.071192 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-81ba-account-create-update-qfjj9" event={"ID":"8ee1f7fe-90fd-4289-b280-f48d0076ce09","Type":"ContainerDied","Data":"e6c44a8797519bcebe5babeb7b0ec7b00111130acc36317f340873c0ad802e8a"} Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.071217 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6c44a8797519bcebe5babeb7b0ec7b00111130acc36317f340873c0ad802e8a" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.077408 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f36e27a-5a90-4f53-8b13-e42e134471f8-kube-api-access-hvbqd" (OuterVolumeSpecName: "kube-api-access-hvbqd") pod "3f36e27a-5a90-4f53-8b13-e42e134471f8" (UID: "3f36e27a-5a90-4f53-8b13-e42e134471f8"). InnerVolumeSpecName "kube-api-access-hvbqd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.081266 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/766cf6f2-1add-4b7e-a5d6-16facb6782dd-kube-api-access-j9b9c" (OuterVolumeSpecName: "kube-api-access-j9b9c") pod "766cf6f2-1add-4b7e-a5d6-16facb6782dd" (UID: "766cf6f2-1add-4b7e-a5d6-16facb6782dd"). InnerVolumeSpecName "kube-api-access-j9b9c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.139663 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-81ba-account-create-update-qfjj9" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.168968 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0405ea9-d37c-40a5-98aa-86ed0d81c288-operator-scripts\") pod \"c0405ea9-d37c-40a5-98aa-86ed0d81c288\" (UID: \"c0405ea9-d37c-40a5-98aa-86ed0d81c288\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.169009 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24qbz\" (UniqueName: \"kubernetes.io/projected/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-kube-api-access-24qbz\") pod \"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5\" (UID: \"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.169046 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjkdd\" (UniqueName: \"kubernetes.io/projected/4ec2c795-db12-41f1-a2cc-536471d01938-kube-api-access-fjkdd\") pod \"4ec2c795-db12-41f1-a2cc-536471d01938\" (UID: \"4ec2c795-db12-41f1-a2cc-536471d01938\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.169067 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7vqc8\" (UniqueName: \"kubernetes.io/projected/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-kube-api-access-7vqc8\") pod \"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679\" (UID: \"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.169110 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b643325f-a8b6-4f76-ad35-52e0c4fd525e-operator-scripts\") pod \"b643325f-a8b6-4f76-ad35-52e0c4fd525e\" (UID: \"b643325f-a8b6-4f76-ad35-52e0c4fd525e\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.169128 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sgqz6\" (UniqueName: \"kubernetes.io/projected/c0405ea9-d37c-40a5-98aa-86ed0d81c288-kube-api-access-sgqz6\") pod \"c0405ea9-d37c-40a5-98aa-86ed0d81c288\" (UID: \"c0405ea9-d37c-40a5-98aa-86ed0d81c288\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.169199 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ec2c795-db12-41f1-a2cc-536471d01938-operator-scripts\") pod \"4ec2c795-db12-41f1-a2cc-536471d01938\" (UID: \"4ec2c795-db12-41f1-a2cc-536471d01938\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.169279 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-operator-scripts\") pod \"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5\" (UID: \"d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.169349 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dlwqt\" (UniqueName: \"kubernetes.io/projected/b643325f-a8b6-4f76-ad35-52e0c4fd525e-kube-api-access-dlwqt\") pod \"b643325f-a8b6-4f76-ad35-52e0c4fd525e\" (UID: \"b643325f-a8b6-4f76-ad35-52e0c4fd525e\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.169368 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-operator-scripts\") pod \"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679\" (UID: \"f43e786a-e9ff-4ca7-a0d6-f9b9ce928679\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.169774 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j9b9c\" (UniqueName: \"kubernetes.io/projected/766cf6f2-1add-4b7e-a5d6-16facb6782dd-kube-api-access-j9b9c\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.169788 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvbqd\" (UniqueName: \"kubernetes.io/projected/3f36e27a-5a90-4f53-8b13-e42e134471f8-kube-api-access-hvbqd\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.170331 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f43e786a-e9ff-4ca7-a0d6-f9b9ce928679" (UID: "f43e786a-e9ff-4ca7-a0d6-f9b9ce928679"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.173443 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c0405ea9-d37c-40a5-98aa-86ed0d81c288-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c0405ea9-d37c-40a5-98aa-86ed0d81c288" (UID: "c0405ea9-d37c-40a5-98aa-86ed0d81c288"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.177594 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b643325f-a8b6-4f76-ad35-52e0c4fd525e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b643325f-a8b6-4f76-ad35-52e0c4fd525e" (UID: "b643325f-a8b6-4f76-ad35-52e0c4fd525e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.177989 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4ec2c795-db12-41f1-a2cc-536471d01938-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4ec2c795-db12-41f1-a2cc-536471d01938" (UID: "4ec2c795-db12-41f1-a2cc-536471d01938"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.178133 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-kube-api-access-24qbz" (OuterVolumeSpecName: "kube-api-access-24qbz") pod "d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5" (UID: "d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5"). InnerVolumeSpecName "kube-api-access-24qbz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.178256 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c0405ea9-d37c-40a5-98aa-86ed0d81c288-kube-api-access-sgqz6" (OuterVolumeSpecName: "kube-api-access-sgqz6") pod "c0405ea9-d37c-40a5-98aa-86ed0d81c288" (UID: "c0405ea9-d37c-40a5-98aa-86ed0d81c288"). InnerVolumeSpecName "kube-api-access-sgqz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.178629 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5" (UID: "d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.181651 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-kube-api-access-7vqc8" (OuterVolumeSpecName: "kube-api-access-7vqc8") pod "f43e786a-e9ff-4ca7-a0d6-f9b9ce928679" (UID: "f43e786a-e9ff-4ca7-a0d6-f9b9ce928679"). InnerVolumeSpecName "kube-api-access-7vqc8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.184605 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b643325f-a8b6-4f76-ad35-52e0c4fd525e-kube-api-access-dlwqt" (OuterVolumeSpecName: "kube-api-access-dlwqt") pod "b643325f-a8b6-4f76-ad35-52e0c4fd525e" (UID: "b643325f-a8b6-4f76-ad35-52e0c4fd525e"). InnerVolumeSpecName "kube-api-access-dlwqt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.203240 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4ec2c795-db12-41f1-a2cc-536471d01938-kube-api-access-fjkdd" (OuterVolumeSpecName: "kube-api-access-fjkdd") pod "4ec2c795-db12-41f1-a2cc-536471d01938" (UID: "4ec2c795-db12-41f1-a2cc-536471d01938"). InnerVolumeSpecName "kube-api-access-fjkdd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.271458 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ee1f7fe-90fd-4289-b280-f48d0076ce09-operator-scripts\") pod \"8ee1f7fe-90fd-4289-b280-f48d0076ce09\" (UID: \"8ee1f7fe-90fd-4289-b280-f48d0076ce09\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.271539 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-82zck\" (UniqueName: \"kubernetes.io/projected/8ee1f7fe-90fd-4289-b280-f48d0076ce09-kube-api-access-82zck\") pod \"8ee1f7fe-90fd-4289-b280-f48d0076ce09\" (UID: \"8ee1f7fe-90fd-4289-b280-f48d0076ce09\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.272069 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.272081 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dlwqt\" (UniqueName: \"kubernetes.io/projected/b643325f-a8b6-4f76-ad35-52e0c4fd525e-kube-api-access-dlwqt\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.272091 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.272100 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c0405ea9-d37c-40a5-98aa-86ed0d81c288-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.272110 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24qbz\" (UniqueName: \"kubernetes.io/projected/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5-kube-api-access-24qbz\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.272121 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjkdd\" (UniqueName: \"kubernetes.io/projected/4ec2c795-db12-41f1-a2cc-536471d01938-kube-api-access-fjkdd\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.272129 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7vqc8\" (UniqueName: \"kubernetes.io/projected/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679-kube-api-access-7vqc8\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.272138 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b643325f-a8b6-4f76-ad35-52e0c4fd525e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.272146 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sgqz6\" (UniqueName: \"kubernetes.io/projected/c0405ea9-d37c-40a5-98aa-86ed0d81c288-kube-api-access-sgqz6\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.272154 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4ec2c795-db12-41f1-a2cc-536471d01938-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.272695 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8ee1f7fe-90fd-4289-b280-f48d0076ce09-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8ee1f7fe-90fd-4289-b280-f48d0076ce09" (UID: "8ee1f7fe-90fd-4289-b280-f48d0076ce09"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.276808 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ee1f7fe-90fd-4289-b280-f48d0076ce09-kube-api-access-82zck" (OuterVolumeSpecName: "kube-api-access-82zck") pod "8ee1f7fe-90fd-4289-b280-f48d0076ce09" (UID: "8ee1f7fe-90fd-4289-b280-f48d0076ce09"). InnerVolumeSpecName "kube-api-access-82zck". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.288632 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.373670 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-dns-svc\") pod \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.373748 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-sb\") pod \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.373818 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-config\") pod \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.373939 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-nb\") pod \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.373964 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgdkf\" (UniqueName: \"kubernetes.io/projected/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-kube-api-access-bgdkf\") pod \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\" (UID: \"d0bc011b-984e-4e56-a3c2-b5ea04a74a73\") " Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.374343 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8ee1f7fe-90fd-4289-b280-f48d0076ce09-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.374355 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-82zck\" (UniqueName: \"kubernetes.io/projected/8ee1f7fe-90fd-4289-b280-f48d0076ce09-kube-api-access-82zck\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.397575 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-kube-api-access-bgdkf" (OuterVolumeSpecName: "kube-api-access-bgdkf") pod "d0bc011b-984e-4e56-a3c2-b5ea04a74a73" (UID: "d0bc011b-984e-4e56-a3c2-b5ea04a74a73"). InnerVolumeSpecName "kube-api-access-bgdkf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.427316 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d0bc011b-984e-4e56-a3c2-b5ea04a74a73" (UID: "d0bc011b-984e-4e56-a3c2-b5ea04a74a73"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.429063 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d0bc011b-984e-4e56-a3c2-b5ea04a74a73" (UID: "d0bc011b-984e-4e56-a3c2-b5ea04a74a73"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.435701 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-config" (OuterVolumeSpecName: "config") pod "d0bc011b-984e-4e56-a3c2-b5ea04a74a73" (UID: "d0bc011b-984e-4e56-a3c2-b5ea04a74a73"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.451332 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d0bc011b-984e-4e56-a3c2-b5ea04a74a73" (UID: "d0bc011b-984e-4e56-a3c2-b5ea04a74a73"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.478454 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.478484 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.478496 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.478507 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:43 crc kubenswrapper[4787]: I1203 17:35:43.478521 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgdkf\" (UniqueName: \"kubernetes.io/projected/d0bc011b-984e-4e56-a3c2-b5ea04a74a73-kube-api-access-bgdkf\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.087029 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-698758b865-kspp6" event={"ID":"d0bc011b-984e-4e56-a3c2-b5ea04a74a73","Type":"ContainerDied","Data":"758be19b486b87d031a3e0e6a590e33442a66ee4a068cd0d71b9442c04030e65"} Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.087093 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-698758b865-kspp6" Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.087323 4787 scope.go:117] "RemoveContainer" containerID="783be89bd64ad820febf63d687389dcb5e50ed3b1d5295a09290c8de26b99293" Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.089571 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-sx9j7" event={"ID":"dadfdc4d-30a6-4edd-88a4-8835c6e433c0","Type":"ContainerStarted","Data":"9f237298f174bca5da3b59c792602109df5fa64d6589447b51dba505bdf71de7"} Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.095986 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-81ba-account-create-update-qfjj9" Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.096417 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7e89b75-951a-44b1-9362-fa52750d7450","Type":"ContainerStarted","Data":"8bfc6705f132cb3739648a9955edbc44e897b09d82ccf4e968c01cfcf1cebce6"} Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.096468 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-d920-account-create-update-nqrk5" Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.096480 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-8e2a-account-create-update-9z9kk" Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.130222 4787 scope.go:117] "RemoveContainer" containerID="78cf7603e71e6836acd703c0ceb8e94dbaa3c540d9cd1520f9b83839afd59c3c" Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.163013 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-sx9j7" podStartSLOduration=3.096849214 podStartE2EDuration="8.162991282s" podCreationTimestamp="2025-12-03 17:35:36 +0000 UTC" firstStartedPulling="2025-12-03 17:35:37.623786047 +0000 UTC m=+1354.441257006" lastFinishedPulling="2025-12-03 17:35:42.689928115 +0000 UTC m=+1359.507399074" observedRunningTime="2025-12-03 17:35:44.146783491 +0000 UTC m=+1360.964254450" watchObservedRunningTime="2025-12-03 17:35:44.162991282 +0000 UTC m=+1360.980462241" Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.223070 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kspp6"] Dec 03 17:35:44 crc kubenswrapper[4787]: I1203 17:35:44.234143 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-698758b865-kspp6"] Dec 03 17:35:45 crc kubenswrapper[4787]: I1203 17:35:45.781841 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d0bc011b-984e-4e56-a3c2-b5ea04a74a73" path="/var/lib/kubelet/pods/d0bc011b-984e-4e56-a3c2-b5ea04a74a73/volumes" Dec 03 17:35:46 crc kubenswrapper[4787]: I1203 17:35:46.119388 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7e89b75-951a-44b1-9362-fa52750d7450","Type":"ContainerStarted","Data":"459631df138478599296d1a7e6542c4fc0fce123cf2ae69052fc2059aa7f9873"} Dec 03 17:35:47 crc kubenswrapper[4787]: I1203 17:35:47.130452 4787 generic.go:334] "Generic (PLEG): container finished" podID="dadfdc4d-30a6-4edd-88a4-8835c6e433c0" containerID="9f237298f174bca5da3b59c792602109df5fa64d6589447b51dba505bdf71de7" exitCode=0 Dec 03 17:35:47 crc kubenswrapper[4787]: I1203 17:35:47.130490 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-sx9j7" event={"ID":"dadfdc4d-30a6-4edd-88a4-8835c6e433c0","Type":"ContainerDied","Data":"9f237298f174bca5da3b59c792602109df5fa64d6589447b51dba505bdf71de7"} Dec 03 17:35:47 crc kubenswrapper[4787]: I1203 17:35:47.133755 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/prometheus-metric-storage-0" event={"ID":"b7e89b75-951a-44b1-9362-fa52750d7450","Type":"ContainerStarted","Data":"369d60765fe53cc690e5ed9288a830ef2696686b7e1cafb31ec31526c23f890e"} Dec 03 17:35:47 crc kubenswrapper[4787]: I1203 17:35:47.178162 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/prometheus-metric-storage-0" podStartSLOduration=16.178144548 podStartE2EDuration="16.178144548s" podCreationTimestamp="2025-12-03 17:35:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:35:47.173986777 +0000 UTC m=+1363.991457746" watchObservedRunningTime="2025-12-03 17:35:47.178144548 +0000 UTC m=+1363.995615507" Dec 03 17:35:48 crc kubenswrapper[4787]: I1203 17:35:48.049216 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-698758b865-kspp6" podUID="d0bc011b-984e-4e56-a3c2-b5ea04a74a73" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.140:5353: i/o timeout" Dec 03 17:35:48 crc kubenswrapper[4787]: I1203 17:35:48.578000 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:48 crc kubenswrapper[4787]: I1203 17:35:48.673449 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8z454\" (UniqueName: \"kubernetes.io/projected/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-kube-api-access-8z454\") pod \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " Dec 03 17:35:48 crc kubenswrapper[4787]: I1203 17:35:48.673563 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-combined-ca-bundle\") pod \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " Dec 03 17:35:48 crc kubenswrapper[4787]: I1203 17:35:48.673660 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-config-data\") pod \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\" (UID: \"dadfdc4d-30a6-4edd-88a4-8835c6e433c0\") " Dec 03 17:35:48 crc kubenswrapper[4787]: I1203 17:35:48.678989 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-kube-api-access-8z454" (OuterVolumeSpecName: "kube-api-access-8z454") pod "dadfdc4d-30a6-4edd-88a4-8835c6e433c0" (UID: "dadfdc4d-30a6-4edd-88a4-8835c6e433c0"). InnerVolumeSpecName "kube-api-access-8z454". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:48 crc kubenswrapper[4787]: I1203 17:35:48.700869 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dadfdc4d-30a6-4edd-88a4-8835c6e433c0" (UID: "dadfdc4d-30a6-4edd-88a4-8835c6e433c0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:48 crc kubenswrapper[4787]: I1203 17:35:48.736316 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-config-data" (OuterVolumeSpecName: "config-data") pod "dadfdc4d-30a6-4edd-88a4-8835c6e433c0" (UID: "dadfdc4d-30a6-4edd-88a4-8835c6e433c0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:35:48 crc kubenswrapper[4787]: I1203 17:35:48.775741 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:48 crc kubenswrapper[4787]: I1203 17:35:48.775784 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:48 crc kubenswrapper[4787]: I1203 17:35:48.775799 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8z454\" (UniqueName: \"kubernetes.io/projected/dadfdc4d-30a6-4edd-88a4-8835c6e433c0-kube-api-access-8z454\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.153105 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-sx9j7" event={"ID":"dadfdc4d-30a6-4edd-88a4-8835c6e433c0","Type":"ContainerDied","Data":"3803d93b643cfb4b576bfc92a9ac70d95a28aeebfbc94c87800bb95af390939e"} Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.153147 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3803d93b643cfb4b576bfc92a9ac70d95a28aeebfbc94c87800bb95af390939e" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.153191 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-sx9j7" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.485414 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8kzjf"] Dec 03 17:35:49 crc kubenswrapper[4787]: E1203 17:35:49.486212 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dadfdc4d-30a6-4edd-88a4-8835c6e433c0" containerName="keystone-db-sync" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486240 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="dadfdc4d-30a6-4edd-88a4-8835c6e433c0" containerName="keystone-db-sync" Dec 03 17:35:49 crc kubenswrapper[4787]: E1203 17:35:49.486259 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ee1f7fe-90fd-4289-b280-f48d0076ce09" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486268 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ee1f7fe-90fd-4289-b280-f48d0076ce09" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: E1203 17:35:49.486289 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="766cf6f2-1add-4b7e-a5d6-16facb6782dd" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486298 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="766cf6f2-1add-4b7e-a5d6-16facb6782dd" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: E1203 17:35:49.486310 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f36e27a-5a90-4f53-8b13-e42e134471f8" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486317 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f36e27a-5a90-4f53-8b13-e42e134471f8" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: E1203 17:35:49.486329 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f43e786a-e9ff-4ca7-a0d6-f9b9ce928679" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486337 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f43e786a-e9ff-4ca7-a0d6-f9b9ce928679" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: E1203 17:35:49.486356 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b643325f-a8b6-4f76-ad35-52e0c4fd525e" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486363 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b643325f-a8b6-4f76-ad35-52e0c4fd525e" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: E1203 17:35:49.486380 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4ec2c795-db12-41f1-a2cc-536471d01938" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486387 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4ec2c795-db12-41f1-a2cc-536471d01938" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: E1203 17:35:49.486427 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0bc011b-984e-4e56-a3c2-b5ea04a74a73" containerName="dnsmasq-dns" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486436 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0bc011b-984e-4e56-a3c2-b5ea04a74a73" containerName="dnsmasq-dns" Dec 03 17:35:49 crc kubenswrapper[4787]: E1203 17:35:49.486454 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c0405ea9-d37c-40a5-98aa-86ed0d81c288" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486463 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c0405ea9-d37c-40a5-98aa-86ed0d81c288" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: E1203 17:35:49.486477 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d0bc011b-984e-4e56-a3c2-b5ea04a74a73" containerName="init" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486485 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d0bc011b-984e-4e56-a3c2-b5ea04a74a73" containerName="init" Dec 03 17:35:49 crc kubenswrapper[4787]: E1203 17:35:49.486499 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486508 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486742 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b643325f-a8b6-4f76-ad35-52e0c4fd525e" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486763 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f36e27a-5a90-4f53-8b13-e42e134471f8" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486779 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f43e786a-e9ff-4ca7-a0d6-f9b9ce928679" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486793 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d0bc011b-984e-4e56-a3c2-b5ea04a74a73" containerName="dnsmasq-dns" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486808 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="766cf6f2-1add-4b7e-a5d6-16facb6782dd" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486822 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4ec2c795-db12-41f1-a2cc-536471d01938" containerName="mariadb-database-create" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486840 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="dadfdc4d-30a6-4edd-88a4-8835c6e433c0" containerName="keystone-db-sync" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486854 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486868 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c0405ea9-d37c-40a5-98aa-86ed0d81c288" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.486887 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ee1f7fe-90fd-4289-b280-f48d0076ce09" containerName="mariadb-account-create-update" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.501334 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.513138 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-m8kgt"] Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.514550 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.524189 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8kzjf"] Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.533902 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-m8kgt"] Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.558218 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.558510 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.558683 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-l49xh" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.559889 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.560004 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597208 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597248 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-config\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597298 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597326 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97kxk\" (UniqueName: \"kubernetes.io/projected/6b821380-6139-4743-8b40-f71ebccc6756-kube-api-access-97kxk\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597356 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597375 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pc5lx\" (UniqueName: \"kubernetes.io/projected/8a2a93be-3b61-4628-a47a-68b85b10b839-kube-api-access-pc5lx\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597405 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-svc\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597438 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-scripts\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597467 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-fernet-keys\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597494 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-config-data\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597527 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-credential-keys\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.597558 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-combined-ca-bundle\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.627812 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-6pgw8"] Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.630600 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6pgw8" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.635800 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-wn2sj" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.647170 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.670923 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-6pgw8"] Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699497 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-svc\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699559 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-scripts\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699585 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-fernet-keys\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699601 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-config-data\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699628 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-config-data\") pod \"heat-db-sync-6pgw8\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " pod="openstack/heat-db-sync-6pgw8" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699650 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-credential-keys\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699690 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-combined-ca-bundle\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699721 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699740 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-config\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699763 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-combined-ca-bundle\") pod \"heat-db-sync-6pgw8\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " pod="openstack/heat-db-sync-6pgw8" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699790 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-px6sz\" (UniqueName: \"kubernetes.io/projected/5e5395cf-a94f-4125-8d92-54b7cb9da3be-kube-api-access-px6sz\") pod \"heat-db-sync-6pgw8\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " pod="openstack/heat-db-sync-6pgw8" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699829 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699850 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97kxk\" (UniqueName: \"kubernetes.io/projected/6b821380-6139-4743-8b40-f71ebccc6756-kube-api-access-97kxk\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699876 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.699894 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pc5lx\" (UniqueName: \"kubernetes.io/projected/8a2a93be-3b61-4628-a47a-68b85b10b839-kube-api-access-pc5lx\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.700966 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-svc\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.706087 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-nb\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.709991 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-config\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.711079 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-swift-storage-0\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.711779 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-sb\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.716967 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-scripts\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.719337 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-credential-keys\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.722250 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-combined-ca-bundle\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.722555 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-fernet-keys\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.723122 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-config-data\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.747743 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pc5lx\" (UniqueName: \"kubernetes.io/projected/8a2a93be-3b61-4628-a47a-68b85b10b839-kube-api-access-pc5lx\") pod \"dnsmasq-dns-847c4cc679-8kzjf\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.766127 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97kxk\" (UniqueName: \"kubernetes.io/projected/6b821380-6139-4743-8b40-f71ebccc6756-kube-api-access-97kxk\") pod \"keystone-bootstrap-m8kgt\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.807051 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-config-data\") pod \"heat-db-sync-6pgw8\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " pod="openstack/heat-db-sync-6pgw8" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.807132 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-combined-ca-bundle\") pod \"heat-db-sync-6pgw8\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " pod="openstack/heat-db-sync-6pgw8" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.807156 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-px6sz\" (UniqueName: \"kubernetes.io/projected/5e5395cf-a94f-4125-8d92-54b7cb9da3be-kube-api-access-px6sz\") pod \"heat-db-sync-6pgw8\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " pod="openstack/heat-db-sync-6pgw8" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.820244 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-config-data\") pod \"heat-db-sync-6pgw8\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " pod="openstack/heat-db-sync-6pgw8" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.826960 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-combined-ca-bundle\") pod \"heat-db-sync-6pgw8\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " pod="openstack/heat-db-sync-6pgw8" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.834824 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.856916 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.906568 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-px6sz\" (UniqueName: \"kubernetes.io/projected/5e5395cf-a94f-4125-8d92-54b7cb9da3be-kube-api-access-px6sz\") pod \"heat-db-sync-6pgw8\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " pod="openstack/heat-db-sync-6pgw8" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.906613 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-qqt8b"] Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.919808 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qqt8b"] Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.919911 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.921822 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.922155 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-j5qpm" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.929067 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 17:35:49 crc kubenswrapper[4787]: I1203 17:35:49.959422 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6pgw8" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:49.999671 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-zmx58"] Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.001598 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.017621 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-combined-ca-bundle\") pod \"neutron-db-sync-qqt8b\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.017717 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7dqpj\" (UniqueName: \"kubernetes.io/projected/af99de84-794c-4fc3-9582-6007e90c71f3-kube-api-access-7dqpj\") pod \"neutron-db-sync-qqt8b\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.017852 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-config\") pod \"neutron-db-sync-qqt8b\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.024088 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-zrgzg"] Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.025361 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.031385 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-zrgzg"] Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.038967 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.039249 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-d8gx4" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.039410 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-zmx58"] Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.039912 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.040101 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-686jj" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.040211 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.054339 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8kzjf"] Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.119093 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-combined-ca-bundle\") pod \"neutron-db-sync-qqt8b\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.119363 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-config-data\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.119492 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-combined-ca-bundle\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.119594 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-scripts\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.119706 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7dqpj\" (UniqueName: \"kubernetes.io/projected/af99de84-794c-4fc3-9582-6007e90c71f3-kube-api-access-7dqpj\") pod \"neutron-db-sync-qqt8b\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.119801 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-db-sync-config-data\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.119935 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-config\") pod \"neutron-db-sync-qqt8b\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.120043 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-db-sync-config-data\") pod \"barbican-db-sync-zrgzg\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.120146 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-combined-ca-bundle\") pod \"barbican-db-sync-zrgzg\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.120370 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-etc-machine-id\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.120482 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk8pj\" (UniqueName: \"kubernetes.io/projected/0308bd88-1943-4aea-a041-b5e7bd9e64f3-kube-api-access-gk8pj\") pod \"barbican-db-sync-zrgzg\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.120577 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwnrs\" (UniqueName: \"kubernetes.io/projected/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-kube-api-access-hwnrs\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.128678 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-combined-ca-bundle\") pod \"neutron-db-sync-qqt8b\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.133109 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-fkrgl"] Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.134675 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.135915 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-config\") pod \"neutron-db-sync-qqt8b\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.155889 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.156865 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.157604 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6nfqd" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.186530 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6xx4r"] Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.192278 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.207210 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7dqpj\" (UniqueName: \"kubernetes.io/projected/af99de84-794c-4fc3-9582-6007e90c71f3-kube-api-access-7dqpj\") pod \"neutron-db-sync-qqt8b\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.221890 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-db-sync-config-data\") pod \"barbican-db-sync-zrgzg\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.221947 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-combined-ca-bundle\") pod \"barbican-db-sync-zrgzg\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.221984 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-etc-machine-id\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.222009 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk8pj\" (UniqueName: \"kubernetes.io/projected/0308bd88-1943-4aea-a041-b5e7bd9e64f3-kube-api-access-gk8pj\") pod \"barbican-db-sync-zrgzg\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.222044 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwnrs\" (UniqueName: \"kubernetes.io/projected/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-kube-api-access-hwnrs\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.222065 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372bef16-5eab-41db-acbf-07274bf3f985-logs\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.222129 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h596f\" (UniqueName: \"kubernetes.io/projected/372bef16-5eab-41db-acbf-07274bf3f985-kube-api-access-h596f\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.222171 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-config-data\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.222189 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-combined-ca-bundle\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.222207 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-scripts\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.222238 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-db-sync-config-data\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.222256 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-scripts\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.222269 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-config-data\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.222287 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-combined-ca-bundle\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.228040 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-etc-machine-id\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.236395 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-db-sync-config-data\") pod \"barbican-db-sync-zrgzg\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.236667 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-combined-ca-bundle\") pod \"barbican-db-sync-zrgzg\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.236913 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-combined-ca-bundle\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.236986 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-db-sync-config-data\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.256868 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-config-data\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.270329 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-scripts\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.271192 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-fkrgl"] Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.275497 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk8pj\" (UniqueName: \"kubernetes.io/projected/0308bd88-1943-4aea-a041-b5e7bd9e64f3-kube-api-access-gk8pj\") pod \"barbican-db-sync-zrgzg\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.286229 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwnrs\" (UniqueName: \"kubernetes.io/projected/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-kube-api-access-hwnrs\") pod \"cinder-db-sync-zmx58\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.302276 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6xx4r"] Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.325407 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.337323 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372bef16-5eab-41db-acbf-07274bf3f985-logs\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.338230 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-config\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.338655 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h596f\" (UniqueName: \"kubernetes.io/projected/372bef16-5eab-41db-acbf-07274bf3f985-kube-api-access-h596f\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.339393 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.340923 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-scripts\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.341886 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-config-data\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.342001 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-combined-ca-bundle\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.342131 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.340548 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.338011 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372bef16-5eab-41db-acbf-07274bf3f985-logs\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.342325 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxg98\" (UniqueName: \"kubernetes.io/projected/43a5db18-9a22-465e-bd9a-4d96572ba17b-kube-api-access-xxg98\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.343928 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.354155 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-scripts\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.360643 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-combined-ca-bundle\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.363501 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.365684 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.368735 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zmx58" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.370093 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-config-data\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.374519 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.374710 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.379941 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h596f\" (UniqueName: \"kubernetes.io/projected/372bef16-5eab-41db-acbf-07274bf3f985-kube-api-access-h596f\") pod \"placement-db-sync-fkrgl\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.409601 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.410132 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.457139 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.457234 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-config\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.457286 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-run-httpd\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.458779 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-config-data\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.458829 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-config\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.458949 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-scripts\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.458972 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.459058 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.459081 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-log-httpd\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.459106 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p7j2\" (UniqueName: \"kubernetes.io/projected/5b18d92e-dfd0-4c8c-bb33-9af70f665417-kube-api-access-7p7j2\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.459173 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.459240 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxg98\" (UniqueName: \"kubernetes.io/projected/43a5db18-9a22-465e-bd9a-4d96572ba17b-kube-api-access-xxg98\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.459265 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.459329 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.459428 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-swift-storage-0\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.460004 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-nb\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.460541 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-svc\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.468210 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-sb\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.481811 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fkrgl" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.497722 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxg98\" (UniqueName: \"kubernetes.io/projected/43a5db18-9a22-465e-bd9a-4d96572ba17b-kube-api-access-xxg98\") pod \"dnsmasq-dns-785d8bcb8c-6xx4r\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.573709 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.573899 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-run-httpd\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.573938 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-config-data\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.574062 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-scripts\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.574085 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.574150 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-log-httpd\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.574168 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p7j2\" (UniqueName: \"kubernetes.io/projected/5b18d92e-dfd0-4c8c-bb33-9af70f665417-kube-api-access-7p7j2\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.574521 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-run-httpd\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.575444 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.576810 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-log-httpd\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.578851 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.579914 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-config-data\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.585162 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.591615 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-scripts\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.609857 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p7j2\" (UniqueName: \"kubernetes.io/projected/5b18d92e-dfd0-4c8c-bb33-9af70f665417-kube-api-access-7p7j2\") pod \"ceilometer-0\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.740134 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:35:50 crc kubenswrapper[4787]: I1203 17:35:50.899596 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-m8kgt"] Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.007098 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-6pgw8"] Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.013152 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8kzjf"] Dec 03 17:35:51 crc kubenswrapper[4787]: W1203 17:35:51.029113 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5e5395cf_a94f_4125_8d92_54b7cb9da3be.slice/crio-b4c95243415c5cb41025f51a0fb32ff227cccba5e5c87a758b7f542accf4234c WatchSource:0}: Error finding container b4c95243415c5cb41025f51a0fb32ff227cccba5e5c87a758b7f542accf4234c: Status 404 returned error can't find the container with id b4c95243415c5cb41025f51a0fb32ff227cccba5e5c87a758b7f542accf4234c Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.211530 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m8kgt" event={"ID":"6b821380-6139-4743-8b40-f71ebccc6756","Type":"ContainerStarted","Data":"efbc30d5352f0b68539dcfdb6ddff04d64f180b15fccf38745eb784081ca9932"} Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.211957 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m8kgt" event={"ID":"6b821380-6139-4743-8b40-f71ebccc6756","Type":"ContainerStarted","Data":"79fe717bb77db876f0a4ebe833d8e9d21bc846fa61c881478effda5d54157cf3"} Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.223481 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6pgw8" event={"ID":"5e5395cf-a94f-4125-8d92-54b7cb9da3be","Type":"ContainerStarted","Data":"b4c95243415c5cb41025f51a0fb32ff227cccba5e5c87a758b7f542accf4234c"} Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.233411 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" event={"ID":"8a2a93be-3b61-4628-a47a-68b85b10b839","Type":"ContainerStarted","Data":"a1bab52ad10055f65d0e6dc34115ff604aa60f3dc98f199c77abaa1b70b6f416"} Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.244075 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-qqt8b"] Dec 03 17:35:51 crc kubenswrapper[4787]: W1203 17:35:51.247260 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcfab6daa_4ee6_4b9a_b60c_4a553cbe4d90.slice/crio-8eb389a049d093a2f6d5ddddd9607bfbf99b8323d31d79fbf9740d50db91f927 WatchSource:0}: Error finding container 8eb389a049d093a2f6d5ddddd9607bfbf99b8323d31d79fbf9740d50db91f927: Status 404 returned error can't find the container with id 8eb389a049d093a2f6d5ddddd9607bfbf99b8323d31d79fbf9740d50db91f927 Dec 03 17:35:51 crc kubenswrapper[4787]: W1203 17:35:51.249870 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podaf99de84_794c_4fc3_9582_6007e90c71f3.slice/crio-bc4d31bc7d4ffebf8453b65d2c9aec50ffece1c03c01dd56d5f3a60b0933295a WatchSource:0}: Error finding container bc4d31bc7d4ffebf8453b65d2c9aec50ffece1c03c01dd56d5f3a60b0933295a: Status 404 returned error can't find the container with id bc4d31bc7d4ffebf8453b65d2c9aec50ffece1c03c01dd56d5f3a60b0933295a Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.278185 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-zmx58"] Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.292845 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-m8kgt" podStartSLOduration=2.292825678 podStartE2EDuration="2.292825678s" podCreationTimestamp="2025-12-03 17:35:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:35:51.235522645 +0000 UTC m=+1368.052993604" watchObservedRunningTime="2025-12-03 17:35:51.292825678 +0000 UTC m=+1368.110296637" Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.573209 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/prometheus-metric-storage-0" Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.692523 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-fkrgl"] Dec 03 17:35:51 crc kubenswrapper[4787]: W1203 17:35:51.703460 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod372bef16_5eab_41db_acbf_07274bf3f985.slice/crio-0cdf1e4bab7d994495e291f6954221ac6e48773ac0c0ed9b359b30af7114616c WatchSource:0}: Error finding container 0cdf1e4bab7d994495e291f6954221ac6e48773ac0c0ed9b359b30af7114616c: Status 404 returned error can't find the container with id 0cdf1e4bab7d994495e291f6954221ac6e48773ac0c0ed9b359b30af7114616c Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.706141 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-zrgzg"] Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.759718 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6xx4r"] Dec 03 17:35:51 crc kubenswrapper[4787]: I1203 17:35:51.810148 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.251773 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b18d92e-dfd0-4c8c-bb33-9af70f665417","Type":"ContainerStarted","Data":"e92f9ea2a393f913af1aaa874d1eff0a6e4dbf1da5675d47e970bf5c1d033484"} Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.259220 4787 generic.go:334] "Generic (PLEG): container finished" podID="8a2a93be-3b61-4628-a47a-68b85b10b839" containerID="ba8cf2647248b321198b2ef5013e2805f8042d4ebd1bef6c869bd6e5cb4d3ef5" exitCode=0 Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.259463 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" event={"ID":"8a2a93be-3b61-4628-a47a-68b85b10b839","Type":"ContainerDied","Data":"ba8cf2647248b321198b2ef5013e2805f8042d4ebd1bef6c869bd6e5cb4d3ef5"} Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.284796 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zmx58" event={"ID":"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90","Type":"ContainerStarted","Data":"8eb389a049d093a2f6d5ddddd9607bfbf99b8323d31d79fbf9740d50db91f927"} Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.292514 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qqt8b" event={"ID":"af99de84-794c-4fc3-9582-6007e90c71f3","Type":"ContainerStarted","Data":"5ea8872fcab9bad9d38cfa5fe6bf0b6a99b91830dbdf9c5b24932474ef68a99f"} Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.292578 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qqt8b" event={"ID":"af99de84-794c-4fc3-9582-6007e90c71f3","Type":"ContainerStarted","Data":"bc4d31bc7d4ffebf8453b65d2c9aec50ffece1c03c01dd56d5f3a60b0933295a"} Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.297061 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fkrgl" event={"ID":"372bef16-5eab-41db-acbf-07274bf3f985","Type":"ContainerStarted","Data":"0cdf1e4bab7d994495e291f6954221ac6e48773ac0c0ed9b359b30af7114616c"} Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.307468 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" event={"ID":"43a5db18-9a22-465e-bd9a-4d96572ba17b","Type":"ContainerStarted","Data":"48dee910c038cebfbb50fdef997fe49ff6ba65ba9b3c3a5cc9b8e57d76b4f43e"} Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.328323 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zrgzg" event={"ID":"0308bd88-1943-4aea-a041-b5e7bd9e64f3","Type":"ContainerStarted","Data":"8f8725812c54d91597a5ac9807e73889846a2f2c18964ad2783b81a975a4f624"} Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.334194 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-qqt8b" podStartSLOduration=3.334173805 podStartE2EDuration="3.334173805s" podCreationTimestamp="2025-12-03 17:35:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:35:52.328410712 +0000 UTC m=+1369.145881671" watchObservedRunningTime="2025-12-03 17:35:52.334173805 +0000 UTC m=+1369.151644764" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.684903 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.837982 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-sb\") pod \"8a2a93be-3b61-4628-a47a-68b85b10b839\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.838233 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-config\") pod \"8a2a93be-3b61-4628-a47a-68b85b10b839\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.838366 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pc5lx\" (UniqueName: \"kubernetes.io/projected/8a2a93be-3b61-4628-a47a-68b85b10b839-kube-api-access-pc5lx\") pod \"8a2a93be-3b61-4628-a47a-68b85b10b839\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.838436 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-nb\") pod \"8a2a93be-3b61-4628-a47a-68b85b10b839\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.838551 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-swift-storage-0\") pod \"8a2a93be-3b61-4628-a47a-68b85b10b839\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.838603 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-svc\") pod \"8a2a93be-3b61-4628-a47a-68b85b10b839\" (UID: \"8a2a93be-3b61-4628-a47a-68b85b10b839\") " Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.851193 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a2a93be-3b61-4628-a47a-68b85b10b839-kube-api-access-pc5lx" (OuterVolumeSpecName: "kube-api-access-pc5lx") pod "8a2a93be-3b61-4628-a47a-68b85b10b839" (UID: "8a2a93be-3b61-4628-a47a-68b85b10b839"). InnerVolumeSpecName "kube-api-access-pc5lx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.899209 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-config" (OuterVolumeSpecName: "config") pod "8a2a93be-3b61-4628-a47a-68b85b10b839" (UID: "8a2a93be-3b61-4628-a47a-68b85b10b839"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.903224 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "8a2a93be-3b61-4628-a47a-68b85b10b839" (UID: "8a2a93be-3b61-4628-a47a-68b85b10b839"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.932756 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "8a2a93be-3b61-4628-a47a-68b85b10b839" (UID: "8a2a93be-3b61-4628-a47a-68b85b10b839"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.937888 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "8a2a93be-3b61-4628-a47a-68b85b10b839" (UID: "8a2a93be-3b61-4628-a47a-68b85b10b839"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.945132 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.945162 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pc5lx\" (UniqueName: \"kubernetes.io/projected/8a2a93be-3b61-4628-a47a-68b85b10b839-kube-api-access-pc5lx\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.945172 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.945199 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.945209 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:52 crc kubenswrapper[4787]: I1203 17:35:52.967130 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "8a2a93be-3b61-4628-a47a-68b85b10b839" (UID: "8a2a93be-3b61-4628-a47a-68b85b10b839"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:35:53 crc kubenswrapper[4787]: I1203 17:35:53.047395 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/8a2a93be-3b61-4628-a47a-68b85b10b839-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:35:53 crc kubenswrapper[4787]: I1203 17:35:53.366344 4787 generic.go:334] "Generic (PLEG): container finished" podID="43a5db18-9a22-465e-bd9a-4d96572ba17b" containerID="463d12055eb81946bfdf419f00afe7fb8856391fba24fde484aaedc7a73727a4" exitCode=0 Dec 03 17:35:53 crc kubenswrapper[4787]: I1203 17:35:53.366434 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" event={"ID":"43a5db18-9a22-465e-bd9a-4d96572ba17b","Type":"ContainerDied","Data":"463d12055eb81946bfdf419f00afe7fb8856391fba24fde484aaedc7a73727a4"} Dec 03 17:35:53 crc kubenswrapper[4787]: I1203 17:35:53.410868 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" Dec 03 17:35:53 crc kubenswrapper[4787]: I1203 17:35:53.412620 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-847c4cc679-8kzjf" event={"ID":"8a2a93be-3b61-4628-a47a-68b85b10b839","Type":"ContainerDied","Data":"a1bab52ad10055f65d0e6dc34115ff604aa60f3dc98f199c77abaa1b70b6f416"} Dec 03 17:35:53 crc kubenswrapper[4787]: I1203 17:35:53.412711 4787 scope.go:117] "RemoveContainer" containerID="ba8cf2647248b321198b2ef5013e2805f8042d4ebd1bef6c869bd6e5cb4d3ef5" Dec 03 17:35:53 crc kubenswrapper[4787]: I1203 17:35:53.447809 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:35:53 crc kubenswrapper[4787]: I1203 17:35:53.547468 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8kzjf"] Dec 03 17:35:53 crc kubenswrapper[4787]: I1203 17:35:53.583811 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-847c4cc679-8kzjf"] Dec 03 17:35:53 crc kubenswrapper[4787]: I1203 17:35:53.787228 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a2a93be-3b61-4628-a47a-68b85b10b839" path="/var/lib/kubelet/pods/8a2a93be-3b61-4628-a47a-68b85b10b839/volumes" Dec 03 17:35:54 crc kubenswrapper[4787]: I1203 17:35:54.427871 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" event={"ID":"43a5db18-9a22-465e-bd9a-4d96572ba17b","Type":"ContainerStarted","Data":"239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320"} Dec 03 17:35:54 crc kubenswrapper[4787]: I1203 17:35:54.429217 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:35:54 crc kubenswrapper[4787]: I1203 17:35:54.454471 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" podStartSLOduration=4.4544569769999995 podStartE2EDuration="4.454456977s" podCreationTimestamp="2025-12-03 17:35:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:35:54.451488858 +0000 UTC m=+1371.268959817" watchObservedRunningTime="2025-12-03 17:35:54.454456977 +0000 UTC m=+1371.271927936" Dec 03 17:35:59 crc kubenswrapper[4787]: I1203 17:35:59.478813 4787 generic.go:334] "Generic (PLEG): container finished" podID="6b821380-6139-4743-8b40-f71ebccc6756" containerID="efbc30d5352f0b68539dcfdb6ddff04d64f180b15fccf38745eb784081ca9932" exitCode=0 Dec 03 17:35:59 crc kubenswrapper[4787]: I1203 17:35:59.478905 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m8kgt" event={"ID":"6b821380-6139-4743-8b40-f71ebccc6756","Type":"ContainerDied","Data":"efbc30d5352f0b68539dcfdb6ddff04d64f180b15fccf38745eb784081ca9932"} Dec 03 17:36:00 crc kubenswrapper[4787]: I1203 17:36:00.577272 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:36:00 crc kubenswrapper[4787]: I1203 17:36:00.653324 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-955rm"] Dec 03 17:36:00 crc kubenswrapper[4787]: I1203 17:36:00.653588 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" podUID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerName="dnsmasq-dns" containerID="cri-o://4452da0c12d65077441132fee773247df42066fd552f22e4aec943fcbdc9a13a" gracePeriod=10 Dec 03 17:36:01 crc kubenswrapper[4787]: I1203 17:36:01.507193 4787 generic.go:334] "Generic (PLEG): container finished" podID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerID="4452da0c12d65077441132fee773247df42066fd552f22e4aec943fcbdc9a13a" exitCode=0 Dec 03 17:36:01 crc kubenswrapper[4787]: I1203 17:36:01.507256 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" event={"ID":"ee7b5f10-8364-4019-990d-fd34a4c2eda5","Type":"ContainerDied","Data":"4452da0c12d65077441132fee773247df42066fd552f22e4aec943fcbdc9a13a"} Dec 03 17:36:01 crc kubenswrapper[4787]: I1203 17:36:01.572930 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/prometheus-metric-storage-0" Dec 03 17:36:01 crc kubenswrapper[4787]: I1203 17:36:01.579229 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/prometheus-metric-storage-0" Dec 03 17:36:02 crc kubenswrapper[4787]: I1203 17:36:02.523407 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/prometheus-metric-storage-0" Dec 03 17:36:02 crc kubenswrapper[4787]: I1203 17:36:02.717768 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" podUID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: connect: connection refused" Dec 03 17:36:06 crc kubenswrapper[4787]: E1203 17:36:06.911494 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-placement-api:current-podified" Dec 03 17:36:06 crc kubenswrapper[4787]: E1203 17:36:06.912459 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:placement-db-sync,Image:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/placement,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:placement-dbsync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-h596f,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42482,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-db-sync-fkrgl_openstack(372bef16-5eab-41db-acbf-07274bf3f985): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:36:06 crc kubenswrapper[4787]: E1203 17:36:06.914273 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/placement-db-sync-fkrgl" podUID="372bef16-5eab-41db-acbf-07274bf3f985" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.028265 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.178544 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-scripts\") pod \"6b821380-6139-4743-8b40-f71ebccc6756\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.178671 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-credential-keys\") pod \"6b821380-6139-4743-8b40-f71ebccc6756\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.178726 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97kxk\" (UniqueName: \"kubernetes.io/projected/6b821380-6139-4743-8b40-f71ebccc6756-kube-api-access-97kxk\") pod \"6b821380-6139-4743-8b40-f71ebccc6756\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.178752 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-combined-ca-bundle\") pod \"6b821380-6139-4743-8b40-f71ebccc6756\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.178848 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-config-data\") pod \"6b821380-6139-4743-8b40-f71ebccc6756\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.178907 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-fernet-keys\") pod \"6b821380-6139-4743-8b40-f71ebccc6756\" (UID: \"6b821380-6139-4743-8b40-f71ebccc6756\") " Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.186930 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b821380-6139-4743-8b40-f71ebccc6756-kube-api-access-97kxk" (OuterVolumeSpecName: "kube-api-access-97kxk") pod "6b821380-6139-4743-8b40-f71ebccc6756" (UID: "6b821380-6139-4743-8b40-f71ebccc6756"). InnerVolumeSpecName "kube-api-access-97kxk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.187546 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "6b821380-6139-4743-8b40-f71ebccc6756" (UID: "6b821380-6139-4743-8b40-f71ebccc6756"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.187803 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-scripts" (OuterVolumeSpecName: "scripts") pod "6b821380-6139-4743-8b40-f71ebccc6756" (UID: "6b821380-6139-4743-8b40-f71ebccc6756"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.200718 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "6b821380-6139-4743-8b40-f71ebccc6756" (UID: "6b821380-6139-4743-8b40-f71ebccc6756"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.218175 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b821380-6139-4743-8b40-f71ebccc6756" (UID: "6b821380-6139-4743-8b40-f71ebccc6756"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.236026 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-config-data" (OuterVolumeSpecName: "config-data") pod "6b821380-6139-4743-8b40-f71ebccc6756" (UID: "6b821380-6139-4743-8b40-f71ebccc6756"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.281582 4787 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.281621 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97kxk\" (UniqueName: \"kubernetes.io/projected/6b821380-6139-4743-8b40-f71ebccc6756-kube-api-access-97kxk\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.281634 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.281643 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.281651 4787 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.281661 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b821380-6139-4743-8b40-f71ebccc6756-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.578008 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-m8kgt" Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.578053 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-m8kgt" event={"ID":"6b821380-6139-4743-8b40-f71ebccc6756","Type":"ContainerDied","Data":"79fe717bb77db876f0a4ebe833d8e9d21bc846fa61c881478effda5d54157cf3"} Dec 03 17:36:07 crc kubenswrapper[4787]: I1203 17:36:07.578110 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79fe717bb77db876f0a4ebe833d8e9d21bc846fa61c881478effda5d54157cf3" Dec 03 17:36:07 crc kubenswrapper[4787]: E1203 17:36:07.583377 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"placement-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-placement-api:current-podified\\\"\"" pod="openstack/placement-db-sync-fkrgl" podUID="372bef16-5eab-41db-acbf-07274bf3f985" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.129388 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-m8kgt"] Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.144107 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-m8kgt"] Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.341159 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-2r5xr"] Dec 03 17:36:08 crc kubenswrapper[4787]: E1203 17:36:08.341621 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a2a93be-3b61-4628-a47a-68b85b10b839" containerName="init" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.341641 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a2a93be-3b61-4628-a47a-68b85b10b839" containerName="init" Dec 03 17:36:08 crc kubenswrapper[4787]: E1203 17:36:08.341676 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b821380-6139-4743-8b40-f71ebccc6756" containerName="keystone-bootstrap" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.341683 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b821380-6139-4743-8b40-f71ebccc6756" containerName="keystone-bootstrap" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.341891 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a2a93be-3b61-4628-a47a-68b85b10b839" containerName="init" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.341923 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b821380-6139-4743-8b40-f71ebccc6756" containerName="keystone-bootstrap" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.342568 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.345351 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.349329 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.353823 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.354161 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.354424 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-l49xh" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.359368 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2r5xr"] Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.427159 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-fernet-keys\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.427262 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c6b8p\" (UniqueName: \"kubernetes.io/projected/7a39f222-78f1-4f59-a557-bcfee7f3104a-kube-api-access-c6b8p\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.427324 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-scripts\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.427355 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-credential-keys\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.427375 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-config-data\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.427580 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-combined-ca-bundle\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.529917 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c6b8p\" (UniqueName: \"kubernetes.io/projected/7a39f222-78f1-4f59-a557-bcfee7f3104a-kube-api-access-c6b8p\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.530433 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-scripts\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.531334 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-credential-keys\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.531452 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-config-data\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.531724 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-combined-ca-bundle\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.531880 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-fernet-keys\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.537510 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-credential-keys\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.537700 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-config-data\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.540627 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-scripts\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.542994 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-combined-ca-bundle\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.547962 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-fernet-keys\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.553734 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c6b8p\" (UniqueName: \"kubernetes.io/projected/7a39f222-78f1-4f59-a557-bcfee7f3104a-kube-api-access-c6b8p\") pod \"keystone-bootstrap-2r5xr\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:08 crc kubenswrapper[4787]: I1203 17:36:08.668328 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:09 crc kubenswrapper[4787]: I1203 17:36:09.778603 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b821380-6139-4743-8b40-f71ebccc6756" path="/var/lib/kubelet/pods/6b821380-6139-4743-8b40-f71ebccc6756/volumes" Dec 03 17:36:12 crc kubenswrapper[4787]: I1203 17:36:12.717942 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" podUID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: i/o timeout" Dec 03 17:36:17 crc kubenswrapper[4787]: I1203 17:36:17.719064 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" podUID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: i/o timeout" Dec 03 17:36:17 crc kubenswrapper[4787]: I1203 17:36:17.719696 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:36:17 crc kubenswrapper[4787]: E1203 17:36:17.983257 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Dec 03 17:36:17 crc kubenswrapper[4787]: E1203 17:36:17.983672 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hwnrs,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-zmx58_openstack(cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:36:17 crc kubenswrapper[4787]: E1203 17:36:17.984846 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-zmx58" podUID="cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" Dec 03 17:36:18 crc kubenswrapper[4787]: E1203 17:36:18.110434 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified" Dec 03 17:36:18 crc kubenswrapper[4787]: E1203 17:36:18.110590 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-px6sz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-6pgw8_openstack(5e5395cf-a94f-4125-8d92-54b7cb9da3be): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:36:18 crc kubenswrapper[4787]: E1203 17:36:18.111946 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-6pgw8" podUID="5e5395cf-a94f-4125-8d92-54b7cb9da3be" Dec 03 17:36:18 crc kubenswrapper[4787]: E1203 17:36:18.392346 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified" Dec 03 17:36:18 crc kubenswrapper[4787]: E1203 17:36:18.392511 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:ceilometer-central-agent,Image:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n589h5c6h66ch558h84h57bh54h9fh65fhdch66fh544h5f4hbh5dh66fh77h5cfh59bh54ch58bh94hfch7dh59ch5ddh65bh676h57h665h87h588q,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:scripts,ReadOnly:true,MountPath:/var/lib/openstack/bin,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/openstack/config,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:ceilometer-central-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7p7j2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[/usr/bin/python3 /var/lib/openstack/bin/centralhealth.py],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:300,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ceilometer-0_openstack(5b18d92e-dfd0-4c8c-bb33-9af70f665417): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.515158 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.623353 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-config\") pod \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.623455 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzqvh\" (UniqueName: \"kubernetes.io/projected/ee7b5f10-8364-4019-990d-fd34a4c2eda5-kube-api-access-bzqvh\") pod \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.623521 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-svc\") pod \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.623566 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-nb\") pod \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.623655 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-sb\") pod \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.623746 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-swift-storage-0\") pod \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\" (UID: \"ee7b5f10-8364-4019-990d-fd34a4c2eda5\") " Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.628988 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee7b5f10-8364-4019-990d-fd34a4c2eda5-kube-api-access-bzqvh" (OuterVolumeSpecName: "kube-api-access-bzqvh") pod "ee7b5f10-8364-4019-990d-fd34a4c2eda5" (UID: "ee7b5f10-8364-4019-990d-fd34a4c2eda5"). InnerVolumeSpecName "kube-api-access-bzqvh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.687767 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "ee7b5f10-8364-4019-990d-fd34a4c2eda5" (UID: "ee7b5f10-8364-4019-990d-fd34a4c2eda5"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.698451 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ee7b5f10-8364-4019-990d-fd34a4c2eda5" (UID: "ee7b5f10-8364-4019-990d-fd34a4c2eda5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.706784 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" event={"ID":"ee7b5f10-8364-4019-990d-fd34a4c2eda5","Type":"ContainerDied","Data":"fb73c9fa36acee53f9df14c114376742eec51f9f1a7a387739e65beff49f6264"} Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.706859 4787 scope.go:117] "RemoveContainer" containerID="4452da0c12d65077441132fee773247df42066fd552f22e4aec943fcbdc9a13a" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.707067 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ee7b5f10-8364-4019-990d-fd34a4c2eda5" (UID: "ee7b5f10-8364-4019-990d-fd34a4c2eda5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.707111 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" Dec 03 17:36:18 crc kubenswrapper[4787]: E1203 17:36:18.717643 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified\\\"\"" pod="openstack/heat-db-sync-6pgw8" podUID="5e5395cf-a94f-4125-8d92-54b7cb9da3be" Dec 03 17:36:18 crc kubenswrapper[4787]: E1203 17:36:18.718891 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-zmx58" podUID="cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.729524 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.729544 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.729554 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzqvh\" (UniqueName: \"kubernetes.io/projected/ee7b5f10-8364-4019-990d-fd34a4c2eda5-kube-api-access-bzqvh\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.729564 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.731271 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ee7b5f10-8364-4019-990d-fd34a4c2eda5" (UID: "ee7b5f10-8364-4019-990d-fd34a4c2eda5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.745525 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-config" (OuterVolumeSpecName: "config") pod "ee7b5f10-8364-4019-990d-fd34a4c2eda5" (UID: "ee7b5f10-8364-4019-990d-fd34a4c2eda5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.778962 4787 scope.go:117] "RemoveContainer" containerID="e24e83892e4407deb827e43d8dca0948a49be43613280b3fa68ad23b347e6a58" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.831659 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.831700 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ee7b5f10-8364-4019-990d-fd34a4c2eda5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:18 crc kubenswrapper[4787]: I1203 17:36:18.923647 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-2r5xr"] Dec 03 17:36:18 crc kubenswrapper[4787]: W1203 17:36:18.928190 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7a39f222_78f1_4f59_a557_bcfee7f3104a.slice/crio-30339699bffa85ade95ca16326ec1fc4a272f0fe5f866299370cdfc19e42a4c0 WatchSource:0}: Error finding container 30339699bffa85ade95ca16326ec1fc4a272f0fe5f866299370cdfc19e42a4c0: Status 404 returned error can't find the container with id 30339699bffa85ade95ca16326ec1fc4a272f0fe5f866299370cdfc19e42a4c0 Dec 03 17:36:19 crc kubenswrapper[4787]: I1203 17:36:19.046913 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-955rm"] Dec 03 17:36:19 crc kubenswrapper[4787]: I1203 17:36:19.058739 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74f6bcbc87-955rm"] Dec 03 17:36:19 crc kubenswrapper[4787]: I1203 17:36:19.715349 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2r5xr" event={"ID":"7a39f222-78f1-4f59-a557-bcfee7f3104a","Type":"ContainerStarted","Data":"59b595c210a822d7a538658e8fea2885d20ffeacd789a83f7a59ca43f7527951"} Dec 03 17:36:19 crc kubenswrapper[4787]: I1203 17:36:19.715670 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2r5xr" event={"ID":"7a39f222-78f1-4f59-a557-bcfee7f3104a","Type":"ContainerStarted","Data":"30339699bffa85ade95ca16326ec1fc4a272f0fe5f866299370cdfc19e42a4c0"} Dec 03 17:36:19 crc kubenswrapper[4787]: I1203 17:36:19.717279 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zrgzg" event={"ID":"0308bd88-1943-4aea-a041-b5e7bd9e64f3","Type":"ContainerStarted","Data":"4e391d1dfb528dbb106ba0a803f077358aa28a2f3ca76df74e46062fc61ffb2b"} Dec 03 17:36:19 crc kubenswrapper[4787]: I1203 17:36:19.750929 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-2r5xr" podStartSLOduration=11.750910168 podStartE2EDuration="11.750910168s" podCreationTimestamp="2025-12-03 17:36:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:36:19.740339408 +0000 UTC m=+1396.557810367" watchObservedRunningTime="2025-12-03 17:36:19.750910168 +0000 UTC m=+1396.568381137" Dec 03 17:36:19 crc kubenswrapper[4787]: I1203 17:36:19.762014 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-zrgzg" podStartSLOduration=4.069648366 podStartE2EDuration="30.761993853s" podCreationTimestamp="2025-12-03 17:35:49 +0000 UTC" firstStartedPulling="2025-12-03 17:35:51.732330325 +0000 UTC m=+1368.549801284" lastFinishedPulling="2025-12-03 17:36:18.424675812 +0000 UTC m=+1395.242146771" observedRunningTime="2025-12-03 17:36:19.757163475 +0000 UTC m=+1396.574634444" watchObservedRunningTime="2025-12-03 17:36:19.761993853 +0000 UTC m=+1396.579464812" Dec 03 17:36:19 crc kubenswrapper[4787]: I1203 17:36:19.776746 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" path="/var/lib/kubelet/pods/ee7b5f10-8364-4019-990d-fd34a4c2eda5/volumes" Dec 03 17:36:20 crc kubenswrapper[4787]: I1203 17:36:20.732551 4787 generic.go:334] "Generic (PLEG): container finished" podID="af99de84-794c-4fc3-9582-6007e90c71f3" containerID="5ea8872fcab9bad9d38cfa5fe6bf0b6a99b91830dbdf9c5b24932474ef68a99f" exitCode=0 Dec 03 17:36:20 crc kubenswrapper[4787]: I1203 17:36:20.732639 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qqt8b" event={"ID":"af99de84-794c-4fc3-9582-6007e90c71f3","Type":"ContainerDied","Data":"5ea8872fcab9bad9d38cfa5fe6bf0b6a99b91830dbdf9c5b24932474ef68a99f"} Dec 03 17:36:20 crc kubenswrapper[4787]: I1203 17:36:20.737318 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b18d92e-dfd0-4c8c-bb33-9af70f665417","Type":"ContainerStarted","Data":"08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38"} Dec 03 17:36:22 crc kubenswrapper[4787]: I1203 17:36:22.720386 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-74f6bcbc87-955rm" podUID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.155:5353: i/o timeout" Dec 03 17:36:22 crc kubenswrapper[4787]: I1203 17:36:22.754948 4787 generic.go:334] "Generic (PLEG): container finished" podID="0308bd88-1943-4aea-a041-b5e7bd9e64f3" containerID="4e391d1dfb528dbb106ba0a803f077358aa28a2f3ca76df74e46062fc61ffb2b" exitCode=0 Dec 03 17:36:22 crc kubenswrapper[4787]: I1203 17:36:22.755030 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zrgzg" event={"ID":"0308bd88-1943-4aea-a041-b5e7bd9e64f3","Type":"ContainerDied","Data":"4e391d1dfb528dbb106ba0a803f077358aa28a2f3ca76df74e46062fc61ffb2b"} Dec 03 17:36:22 crc kubenswrapper[4787]: I1203 17:36:22.757471 4787 generic.go:334] "Generic (PLEG): container finished" podID="7a39f222-78f1-4f59-a557-bcfee7f3104a" containerID="59b595c210a822d7a538658e8fea2885d20ffeacd789a83f7a59ca43f7527951" exitCode=0 Dec 03 17:36:22 crc kubenswrapper[4787]: I1203 17:36:22.757501 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2r5xr" event={"ID":"7a39f222-78f1-4f59-a557-bcfee7f3104a","Type":"ContainerDied","Data":"59b595c210a822d7a538658e8fea2885d20ffeacd789a83f7a59ca43f7527951"} Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.569805 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.625694 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.662245 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.762674 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7dqpj\" (UniqueName: \"kubernetes.io/projected/af99de84-794c-4fc3-9582-6007e90c71f3-kube-api-access-7dqpj\") pod \"af99de84-794c-4fc3-9582-6007e90c71f3\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.763039 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-db-sync-config-data\") pod \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.763136 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-config\") pod \"af99de84-794c-4fc3-9582-6007e90c71f3\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.763158 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-combined-ca-bundle\") pod \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.763182 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk8pj\" (UniqueName: \"kubernetes.io/projected/0308bd88-1943-4aea-a041-b5e7bd9e64f3-kube-api-access-gk8pj\") pod \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\" (UID: \"0308bd88-1943-4aea-a041-b5e7bd9e64f3\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.763405 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-combined-ca-bundle\") pod \"af99de84-794c-4fc3-9582-6007e90c71f3\" (UID: \"af99de84-794c-4fc3-9582-6007e90c71f3\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.763456 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-combined-ca-bundle\") pod \"7a39f222-78f1-4f59-a557-bcfee7f3104a\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.768837 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0308bd88-1943-4aea-a041-b5e7bd9e64f3" (UID: "0308bd88-1943-4aea-a041-b5e7bd9e64f3"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.768868 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0308bd88-1943-4aea-a041-b5e7bd9e64f3-kube-api-access-gk8pj" (OuterVolumeSpecName: "kube-api-access-gk8pj") pod "0308bd88-1943-4aea-a041-b5e7bd9e64f3" (UID: "0308bd88-1943-4aea-a041-b5e7bd9e64f3"). InnerVolumeSpecName "kube-api-access-gk8pj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.768855 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af99de84-794c-4fc3-9582-6007e90c71f3-kube-api-access-7dqpj" (OuterVolumeSpecName: "kube-api-access-7dqpj") pod "af99de84-794c-4fc3-9582-6007e90c71f3" (UID: "af99de84-794c-4fc3-9582-6007e90c71f3"). InnerVolumeSpecName "kube-api-access-7dqpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.783106 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-2r5xr" event={"ID":"7a39f222-78f1-4f59-a557-bcfee7f3104a","Type":"ContainerDied","Data":"30339699bffa85ade95ca16326ec1fc4a272f0fe5f866299370cdfc19e42a4c0"} Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.783151 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30339699bffa85ade95ca16326ec1fc4a272f0fe5f866299370cdfc19e42a4c0" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.783265 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-2r5xr" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.793036 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-zrgzg" event={"ID":"0308bd88-1943-4aea-a041-b5e7bd9e64f3","Type":"ContainerDied","Data":"8f8725812c54d91597a5ac9807e73889846a2f2c18964ad2783b81a975a4f624"} Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.793075 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8f8725812c54d91597a5ac9807e73889846a2f2c18964ad2783b81a975a4f624" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.793130 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-zrgzg" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.799383 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "af99de84-794c-4fc3-9582-6007e90c71f3" (UID: "af99de84-794c-4fc3-9582-6007e90c71f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.806112 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b18d92e-dfd0-4c8c-bb33-9af70f665417","Type":"ContainerStarted","Data":"f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748"} Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.811692 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0308bd88-1943-4aea-a041-b5e7bd9e64f3" (UID: "0308bd88-1943-4aea-a041-b5e7bd9e64f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.815786 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-qqt8b" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.816226 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-qqt8b" event={"ID":"af99de84-794c-4fc3-9582-6007e90c71f3","Type":"ContainerDied","Data":"bc4d31bc7d4ffebf8453b65d2c9aec50ffece1c03c01dd56d5f3a60b0933295a"} Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.816253 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc4d31bc7d4ffebf8453b65d2c9aec50ffece1c03c01dd56d5f3a60b0933295a" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.818488 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fkrgl" event={"ID":"372bef16-5eab-41db-acbf-07274bf3f985","Type":"ContainerStarted","Data":"b9dfc7706b50afd1dd55f3a25a7644cf617bc170fecd695147bb42b76c618dc5"} Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.820477 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7a39f222-78f1-4f59-a557-bcfee7f3104a" (UID: "7a39f222-78f1-4f59-a557-bcfee7f3104a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.826507 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-config" (OuterVolumeSpecName: "config") pod "af99de84-794c-4fc3-9582-6007e90c71f3" (UID: "af99de84-794c-4fc3-9582-6007e90c71f3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.852669 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-fkrgl" podStartSLOduration=2.145423759 podStartE2EDuration="34.852648111s" podCreationTimestamp="2025-12-03 17:35:50 +0000 UTC" firstStartedPulling="2025-12-03 17:35:51.709083017 +0000 UTC m=+1368.526553976" lastFinishedPulling="2025-12-03 17:36:24.416307369 +0000 UTC m=+1401.233778328" observedRunningTime="2025-12-03 17:36:24.844410762 +0000 UTC m=+1401.661881741" watchObservedRunningTime="2025-12-03 17:36:24.852648111 +0000 UTC m=+1401.670119080" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.864313 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-config-data\") pod \"7a39f222-78f1-4f59-a557-bcfee7f3104a\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.864381 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c6b8p\" (UniqueName: \"kubernetes.io/projected/7a39f222-78f1-4f59-a557-bcfee7f3104a-kube-api-access-c6b8p\") pod \"7a39f222-78f1-4f59-a557-bcfee7f3104a\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.865387 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-credential-keys\") pod \"7a39f222-78f1-4f59-a557-bcfee7f3104a\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.865457 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-scripts\") pod \"7a39f222-78f1-4f59-a557-bcfee7f3104a\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.865485 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-fernet-keys\") pod \"7a39f222-78f1-4f59-a557-bcfee7f3104a\" (UID: \"7a39f222-78f1-4f59-a557-bcfee7f3104a\") " Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.865972 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7dqpj\" (UniqueName: \"kubernetes.io/projected/af99de84-794c-4fc3-9582-6007e90c71f3-kube-api-access-7dqpj\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.865991 4787 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.866001 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.866009 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0308bd88-1943-4aea-a041-b5e7bd9e64f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.866031 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk8pj\" (UniqueName: \"kubernetes.io/projected/0308bd88-1943-4aea-a041-b5e7bd9e64f3-kube-api-access-gk8pj\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.866040 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/af99de84-794c-4fc3-9582-6007e90c71f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.866049 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.868002 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "7a39f222-78f1-4f59-a557-bcfee7f3104a" (UID: "7a39f222-78f1-4f59-a557-bcfee7f3104a"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.868729 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a39f222-78f1-4f59-a557-bcfee7f3104a-kube-api-access-c6b8p" (OuterVolumeSpecName: "kube-api-access-c6b8p") pod "7a39f222-78f1-4f59-a557-bcfee7f3104a" (UID: "7a39f222-78f1-4f59-a557-bcfee7f3104a"). InnerVolumeSpecName "kube-api-access-c6b8p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.869629 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-scripts" (OuterVolumeSpecName: "scripts") pod "7a39f222-78f1-4f59-a557-bcfee7f3104a" (UID: "7a39f222-78f1-4f59-a557-bcfee7f3104a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.869653 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "7a39f222-78f1-4f59-a557-bcfee7f3104a" (UID: "7a39f222-78f1-4f59-a557-bcfee7f3104a"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.894884 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-config-data" (OuterVolumeSpecName: "config-data") pod "7a39f222-78f1-4f59-a557-bcfee7f3104a" (UID: "7a39f222-78f1-4f59-a557-bcfee7f3104a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.967647 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.967693 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c6b8p\" (UniqueName: \"kubernetes.io/projected/7a39f222-78f1-4f59-a557-bcfee7f3104a-kube-api-access-c6b8p\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.967706 4787 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.967716 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.967726 4787 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/7a39f222-78f1-4f59-a557-bcfee7f3104a-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.978754 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-69bdc498c7-wqxdv"] Dec 03 17:36:24 crc kubenswrapper[4787]: E1203 17:36:24.979134 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af99de84-794c-4fc3-9582-6007e90c71f3" containerName="neutron-db-sync" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.979150 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="af99de84-794c-4fc3-9582-6007e90c71f3" containerName="neutron-db-sync" Dec 03 17:36:24 crc kubenswrapper[4787]: E1203 17:36:24.979163 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0308bd88-1943-4aea-a041-b5e7bd9e64f3" containerName="barbican-db-sync" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.979170 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="0308bd88-1943-4aea-a041-b5e7bd9e64f3" containerName="barbican-db-sync" Dec 03 17:36:24 crc kubenswrapper[4787]: E1203 17:36:24.979189 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerName="init" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.979196 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerName="init" Dec 03 17:36:24 crc kubenswrapper[4787]: E1203 17:36:24.979209 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a39f222-78f1-4f59-a557-bcfee7f3104a" containerName="keystone-bootstrap" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.979215 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a39f222-78f1-4f59-a557-bcfee7f3104a" containerName="keystone-bootstrap" Dec 03 17:36:24 crc kubenswrapper[4787]: E1203 17:36:24.979222 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerName="dnsmasq-dns" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.979228 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerName="dnsmasq-dns" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.979399 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="0308bd88-1943-4aea-a041-b5e7bd9e64f3" containerName="barbican-db-sync" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.979411 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a39f222-78f1-4f59-a557-bcfee7f3104a" containerName="keystone-bootstrap" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.979424 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee7b5f10-8364-4019-990d-fd34a4c2eda5" containerName="dnsmasq-dns" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.979437 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="af99de84-794c-4fc3-9582-6007e90c71f3" containerName="neutron-db-sync" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.980099 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.996977 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 03 17:36:24 crc kubenswrapper[4787]: I1203 17:36:24.997182 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.021260 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-69bdc498c7-wqxdv"] Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.059076 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5f6df7d97f-5vmdt"] Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.069939 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-internal-tls-certs\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.070038 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-fernet-keys\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.070071 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4cpwh\" (UniqueName: \"kubernetes.io/projected/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-kube-api-access-4cpwh\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.070120 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-public-tls-certs\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.070170 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-config-data\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.070226 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-scripts\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.070340 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-credential-keys\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.070365 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.070392 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-combined-ca-bundle\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.075383 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.083270 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc"] Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.084967 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.092796 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.138289 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5f6df7d97f-5vmdt"] Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176094 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65180fdc-f51b-45e7-ac70-05b9489e1201-combined-ca-bundle\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176133 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65180fdc-f51b-45e7-ac70-05b9489e1201-config-data\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176159 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-combined-ca-bundle\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176173 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83ca165-acb3-4e27-b8f3-519897743134-config-data\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176196 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65180fdc-f51b-45e7-ac70-05b9489e1201-logs\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176213 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhwkm\" (UniqueName: \"kubernetes.io/projected/65180fdc-f51b-45e7-ac70-05b9489e1201-kube-api-access-zhwkm\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176232 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d83ca165-acb3-4e27-b8f3-519897743134-logs\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176247 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bt4p6\" (UniqueName: \"kubernetes.io/projected/d83ca165-acb3-4e27-b8f3-519897743134-kube-api-access-bt4p6\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176264 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-internal-tls-certs\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176295 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-fernet-keys\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176311 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4cpwh\" (UniqueName: \"kubernetes.io/projected/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-kube-api-access-4cpwh\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176327 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-public-tls-certs\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176351 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65180fdc-f51b-45e7-ac70-05b9489e1201-config-data-custom\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176370 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-config-data\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176393 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83ca165-acb3-4e27-b8f3-519897743134-combined-ca-bundle\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176417 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-scripts\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176459 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83ca165-acb3-4e27-b8f3-519897743134-config-data-custom\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.176495 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-credential-keys\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.198343 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-fernet-keys\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.200108 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-public-tls-certs\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.203681 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-scripts\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.204044 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-config-data\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.207617 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-internal-tls-certs\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.207736 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc"] Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.218204 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-credential-keys\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.225651 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4cpwh\" (UniqueName: \"kubernetes.io/projected/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-kube-api-access-4cpwh\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.225798 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c244629d-4ad6-4a6f-9f3d-eaceda01c7e8-combined-ca-bundle\") pod \"keystone-69bdc498c7-wqxdv\" (UID: \"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8\") " pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.245226 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-5cq5p"] Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.251301 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.270661 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-5cq5p"] Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.278442 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65180fdc-f51b-45e7-ac70-05b9489e1201-config-data-custom\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.278497 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83ca165-acb3-4e27-b8f3-519897743134-combined-ca-bundle\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.278561 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83ca165-acb3-4e27-b8f3-519897743134-config-data-custom\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.278610 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65180fdc-f51b-45e7-ac70-05b9489e1201-combined-ca-bundle\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.278628 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65180fdc-f51b-45e7-ac70-05b9489e1201-config-data\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.278709 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83ca165-acb3-4e27-b8f3-519897743134-config-data\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.278755 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65180fdc-f51b-45e7-ac70-05b9489e1201-logs\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.278807 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhwkm\" (UniqueName: \"kubernetes.io/projected/65180fdc-f51b-45e7-ac70-05b9489e1201-kube-api-access-zhwkm\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.278835 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d83ca165-acb3-4e27-b8f3-519897743134-logs\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.278853 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bt4p6\" (UniqueName: \"kubernetes.io/projected/d83ca165-acb3-4e27-b8f3-519897743134-kube-api-access-bt4p6\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.282759 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/65180fdc-f51b-45e7-ac70-05b9489e1201-logs\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.284513 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d83ca165-acb3-4e27-b8f3-519897743134-logs\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.289202 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d83ca165-acb3-4e27-b8f3-519897743134-combined-ca-bundle\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.289941 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/65180fdc-f51b-45e7-ac70-05b9489e1201-config-data-custom\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.294370 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d83ca165-acb3-4e27-b8f3-519897743134-config-data-custom\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.298822 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/65180fdc-f51b-45e7-ac70-05b9489e1201-combined-ca-bundle\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.300547 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-bcffbb96d-q8vwl"] Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.300723 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/65180fdc-f51b-45e7-ac70-05b9489e1201-config-data\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.301276 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d83ca165-acb3-4e27-b8f3-519897743134-config-data\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.301928 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bt4p6\" (UniqueName: \"kubernetes.io/projected/d83ca165-acb3-4e27-b8f3-519897743134-kube-api-access-bt4p6\") pod \"barbican-keystone-listener-78b7c4ccf4-bxfhc\" (UID: \"d83ca165-acb3-4e27-b8f3-519897743134\") " pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.302264 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.304283 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.307989 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhwkm\" (UniqueName: \"kubernetes.io/projected/65180fdc-f51b-45e7-ac70-05b9489e1201-kube-api-access-zhwkm\") pod \"barbican-worker-5f6df7d97f-5vmdt\" (UID: \"65180fdc-f51b-45e7-ac70-05b9489e1201\") " pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.326260 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-bcffbb96d-q8vwl"] Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.341574 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.380219 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-config\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.380260 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.380476 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.380554 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.380815 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.380880 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wpg7\" (UniqueName: \"kubernetes.io/projected/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-kube-api-access-4wpg7\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.432437 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5f6df7d97f-5vmdt" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.482181 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.482229 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data-custom\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.482265 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wpg7\" (UniqueName: \"kubernetes.io/projected/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-kube-api-access-4wpg7\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.482312 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.482328 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e993934f-08d2-40c2-a9af-1b262dc941e9-logs\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.482346 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-combined-ca-bundle\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.482366 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-config\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.482402 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.482420 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qlgp\" (UniqueName: \"kubernetes.io/projected/e993934f-08d2-40c2-a9af-1b262dc941e9-kube-api-access-4qlgp\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.482460 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.482483 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.483011 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-nb\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.483110 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-swift-storage-0\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.483595 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-config\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.484257 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-svc\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.484310 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-sb\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.523445 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wpg7\" (UniqueName: \"kubernetes.io/projected/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-kube-api-access-4wpg7\") pod \"dnsmasq-dns-586bdc5f9-5cq5p\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.536230 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.584210 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data-custom\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.584308 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.584328 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e993934f-08d2-40c2-a9af-1b262dc941e9-logs\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.584551 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-combined-ca-bundle\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.585235 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qlgp\" (UniqueName: \"kubernetes.io/projected/e993934f-08d2-40c2-a9af-1b262dc941e9-kube-api-access-4qlgp\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.585321 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e993934f-08d2-40c2-a9af-1b262dc941e9-logs\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.589174 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.601353 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data-custom\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.601886 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.608567 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qlgp\" (UniqueName: \"kubernetes.io/projected/e993934f-08d2-40c2-a9af-1b262dc941e9-kube-api-access-4qlgp\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.609727 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-combined-ca-bundle\") pod \"barbican-api-bcffbb96d-q8vwl\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:25 crc kubenswrapper[4787]: I1203 17:36:25.650863 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.008923 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-5cq5p"] Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.106117 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-tw5cn"] Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.108043 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.130669 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-69d9b47bb8-6fqn7"] Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.133266 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.144199 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.144275 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.144337 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.144358 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-config\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.144398 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.144437 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wflpf\" (UniqueName: \"kubernetes.io/projected/eedf35c9-ac35-4a92-a1ea-c12beea2588b-kube-api-access-wflpf\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.151683 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-j5qpm" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.151865 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.152090 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.152202 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.209654 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-tw5cn"] Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.236518 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-69bdc498c7-wqxdv"] Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.248324 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-httpd-config\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.248399 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.248427 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-ovndb-tls-certs\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.248451 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j7pb\" (UniqueName: \"kubernetes.io/projected/a5a47435-c445-47ff-8806-aa422090813c-kube-api-access-4j7pb\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.248483 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-config\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.248540 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.248599 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wflpf\" (UniqueName: \"kubernetes.io/projected/eedf35c9-ac35-4a92-a1ea-c12beea2588b-kube-api-access-wflpf\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.248644 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.248701 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-config\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.248733 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.248782 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-combined-ca-bundle\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.250250 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-sb\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.250292 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-swift-storage-0\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.250920 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-config\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.251146 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-nb\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.254974 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-svc\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.263062 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69d9b47bb8-6fqn7"] Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.311453 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wflpf\" (UniqueName: \"kubernetes.io/projected/eedf35c9-ac35-4a92-a1ea-c12beea2588b-kube-api-access-wflpf\") pod \"dnsmasq-dns-85ff748b95-tw5cn\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.350441 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-config\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.350505 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-combined-ca-bundle\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.350537 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-httpd-config\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.350561 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-ovndb-tls-certs\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.350578 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j7pb\" (UniqueName: \"kubernetes.io/projected/a5a47435-c445-47ff-8806-aa422090813c-kube-api-access-4j7pb\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.354817 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-httpd-config\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.362946 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-combined-ca-bundle\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.373823 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-ovndb-tls-certs\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.381627 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-config\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.395192 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j7pb\" (UniqueName: \"kubernetes.io/projected/a5a47435-c445-47ff-8806-aa422090813c-kube-api-access-4j7pb\") pod \"neutron-69d9b47bb8-6fqn7\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: W1203 17:36:26.577126 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod65180fdc_f51b_45e7_ac70_05b9489e1201.slice/crio-c212308897c32a7e4ceac8f92d771054046fc374b253287d1d9b5a45b352edff WatchSource:0}: Error finding container c212308897c32a7e4ceac8f92d771054046fc374b253287d1d9b5a45b352edff: Status 404 returned error can't find the container with id c212308897c32a7e4ceac8f92d771054046fc374b253287d1d9b5a45b352edff Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.587146 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5f6df7d97f-5vmdt"] Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.651048 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.873865 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-bcffbb96d-q8vwl"] Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.891897 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:26 crc kubenswrapper[4787]: I1203 17:36:26.928390 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc"] Dec 03 17:36:27 crc kubenswrapper[4787]: I1203 17:36:27.014546 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcffbb96d-q8vwl" event={"ID":"e993934f-08d2-40c2-a9af-1b262dc941e9","Type":"ContainerStarted","Data":"13f236f815e8d83bc381faeea8856c5a7decde53ba8887420c1ebc83bceaa7d5"} Dec 03 17:36:27 crc kubenswrapper[4787]: I1203 17:36:27.016792 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5f6df7d97f-5vmdt" event={"ID":"65180fdc-f51b-45e7-ac70-05b9489e1201","Type":"ContainerStarted","Data":"c212308897c32a7e4ceac8f92d771054046fc374b253287d1d9b5a45b352edff"} Dec 03 17:36:27 crc kubenswrapper[4787]: I1203 17:36:27.018731 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-69bdc498c7-wqxdv" event={"ID":"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8","Type":"ContainerStarted","Data":"d0552595d2541cf4beff870b0d4c8a9608a89ee243c59662743a899b9c4be7e2"} Dec 03 17:36:27 crc kubenswrapper[4787]: I1203 17:36:27.018767 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-69bdc498c7-wqxdv" event={"ID":"c244629d-4ad6-4a6f-9f3d-eaceda01c7e8","Type":"ContainerStarted","Data":"e8f4ca7f72cfe2582493407ab353443ecead0d376bc813397d2206fa9e7b3a94"} Dec 03 17:36:27 crc kubenswrapper[4787]: I1203 17:36:27.018904 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:27 crc kubenswrapper[4787]: I1203 17:36:27.029320 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" event={"ID":"d83ca165-acb3-4e27-b8f3-519897743134","Type":"ContainerStarted","Data":"17b0beaa216d7cb7a0cd3807d331fea937cc2e7c8f6ddb9aac386ed01c95d9ad"} Dec 03 17:36:27 crc kubenswrapper[4787]: I1203 17:36:27.048704 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-69bdc498c7-wqxdv" podStartSLOduration=3.048682766 podStartE2EDuration="3.048682766s" podCreationTimestamp="2025-12-03 17:36:24 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:36:27.042500861 +0000 UTC m=+1403.859971820" watchObservedRunningTime="2025-12-03 17:36:27.048682766 +0000 UTC m=+1403.866153725" Dec 03 17:36:27 crc kubenswrapper[4787]: I1203 17:36:27.102588 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-5cq5p"] Dec 03 17:36:27 crc kubenswrapper[4787]: W1203 17:36:27.115695 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d1a6f4c_3bf3_4601_b2fe_62373ee28009.slice/crio-8a96b5ae95157247a99881d5fa45722e6932547129a6a4a5214e0b3d8c673e7f WatchSource:0}: Error finding container 8a96b5ae95157247a99881d5fa45722e6932547129a6a4a5214e0b3d8c673e7f: Status 404 returned error can't find the container with id 8a96b5ae95157247a99881d5fa45722e6932547129a6a4a5214e0b3d8c673e7f Dec 03 17:36:27 crc kubenswrapper[4787]: I1203 17:36:27.313110 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-tw5cn"] Dec 03 17:36:27 crc kubenswrapper[4787]: W1203 17:36:27.334289 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeedf35c9_ac35_4a92_a1ea_c12beea2588b.slice/crio-b493e649326147ad0157d611cd34d66154f43d03102d626bb20f6f6461191a88 WatchSource:0}: Error finding container b493e649326147ad0157d611cd34d66154f43d03102d626bb20f6f6461191a88: Status 404 returned error can't find the container with id b493e649326147ad0157d611cd34d66154f43d03102d626bb20f6f6461191a88 Dec 03 17:36:27 crc kubenswrapper[4787]: I1203 17:36:27.733305 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-69d9b47bb8-6fqn7"] Dec 03 17:36:27 crc kubenswrapper[4787]: W1203 17:36:27.787061 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5a47435_c445_47ff_8806_aa422090813c.slice/crio-5aa581fdb2538a587a267514a9d64b67c51264c7b7f1bb48222d64709bf3175d WatchSource:0}: Error finding container 5aa581fdb2538a587a267514a9d64b67c51264c7b7f1bb48222d64709bf3175d: Status 404 returned error can't find the container with id 5aa581fdb2538a587a267514a9d64b67c51264c7b7f1bb48222d64709bf3175d Dec 03 17:36:28 crc kubenswrapper[4787]: I1203 17:36:28.098753 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d9b47bb8-6fqn7" event={"ID":"a5a47435-c445-47ff-8806-aa422090813c","Type":"ContainerStarted","Data":"5aa581fdb2538a587a267514a9d64b67c51264c7b7f1bb48222d64709bf3175d"} Dec 03 17:36:28 crc kubenswrapper[4787]: I1203 17:36:28.100349 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" event={"ID":"eedf35c9-ac35-4a92-a1ea-c12beea2588b","Type":"ContainerStarted","Data":"b493e649326147ad0157d611cd34d66154f43d03102d626bb20f6f6461191a88"} Dec 03 17:36:28 crc kubenswrapper[4787]: I1203 17:36:28.103217 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcffbb96d-q8vwl" event={"ID":"e993934f-08d2-40c2-a9af-1b262dc941e9","Type":"ContainerStarted","Data":"5fc3208f89bdfbfbd0258302400b81497e03a419922d16845c61a526e87a0aef"} Dec 03 17:36:28 crc kubenswrapper[4787]: I1203 17:36:28.110102 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" event={"ID":"7d1a6f4c-3bf3-4601-b2fe-62373ee28009","Type":"ContainerStarted","Data":"8a96b5ae95157247a99881d5fa45722e6932547129a6a4a5214e0b3d8c673e7f"} Dec 03 17:36:28 crc kubenswrapper[4787]: I1203 17:36:28.920034 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-9fd5cd74f-xdkkz"] Dec 03 17:36:28 crc kubenswrapper[4787]: I1203 17:36:28.922980 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:28 crc kubenswrapper[4787]: I1203 17:36:28.928840 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 03 17:36:28 crc kubenswrapper[4787]: I1203 17:36:28.938351 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 03 17:36:28 crc kubenswrapper[4787]: I1203 17:36:28.982535 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9fd5cd74f-xdkkz"] Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.086041 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-public-tls-certs\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.086134 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-config\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.086164 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-ovndb-tls-certs\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.086202 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-combined-ca-bundle\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.086229 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-internal-tls-certs\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.086295 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-httpd-config\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.086338 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcrs9\" (UniqueName: \"kubernetes.io/projected/593e4809-69d6-43bf-ae94-a5d96431ca2f-kube-api-access-tcrs9\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.187747 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-ovndb-tls-certs\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.187836 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-combined-ca-bundle\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.187874 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-internal-tls-certs\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.187959 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-httpd-config\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.188037 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcrs9\" (UniqueName: \"kubernetes.io/projected/593e4809-69d6-43bf-ae94-a5d96431ca2f-kube-api-access-tcrs9\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.188077 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-public-tls-certs\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.188128 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-config\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.192874 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-httpd-config\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.193114 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-combined-ca-bundle\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.193437 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-config\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.202446 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-internal-tls-certs\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.208669 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcrs9\" (UniqueName: \"kubernetes.io/projected/593e4809-69d6-43bf-ae94-a5d96431ca2f-kube-api-access-tcrs9\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.210604 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-public-tls-certs\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.211102 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/593e4809-69d6-43bf-ae94-a5d96431ca2f-ovndb-tls-certs\") pod \"neutron-9fd5cd74f-xdkkz\" (UID: \"593e4809-69d6-43bf-ae94-a5d96431ca2f\") " pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:29 crc kubenswrapper[4787]: I1203 17:36:29.316457 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:30 crc kubenswrapper[4787]: I1203 17:36:30.175158 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-9fd5cd74f-xdkkz"] Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.143323 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9fd5cd74f-xdkkz" event={"ID":"593e4809-69d6-43bf-ae94-a5d96431ca2f","Type":"ContainerStarted","Data":"db333bf37415f17295a52f4140fd11ad2ea4f358f2ba76ae0ff3e1b7ee9b82cc"} Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.823606 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-68677f5b6b-9rwg7"] Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.825200 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.827333 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.829261 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.839783 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-68677f5b6b-9rwg7"] Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.949449 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-public-tls-certs\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.949566 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlzl2\" (UniqueName: \"kubernetes.io/projected/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-kube-api-access-vlzl2\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.949670 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-config-data\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.949730 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-internal-tls-certs\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.949783 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-config-data-custom\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.949893 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-combined-ca-bundle\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:31 crc kubenswrapper[4787]: I1203 17:36:31.949953 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-logs\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.051672 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-internal-tls-certs\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.052128 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-config-data-custom\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.052228 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-combined-ca-bundle\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.052268 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-logs\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.052314 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-public-tls-certs\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.052353 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlzl2\" (UniqueName: \"kubernetes.io/projected/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-kube-api-access-vlzl2\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.052432 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-config-data\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.053270 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-logs\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.058429 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-public-tls-certs\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.058439 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-combined-ca-bundle\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.058948 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-internal-tls-certs\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.059789 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-config-data\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.060789 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-config-data-custom\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.073172 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlzl2\" (UniqueName: \"kubernetes.io/projected/9dc4396d-2f35-42e0-a406-4f0a112fe1cd-kube-api-access-vlzl2\") pod \"barbican-api-68677f5b6b-9rwg7\" (UID: \"9dc4396d-2f35-42e0-a406-4f0a112fe1cd\") " pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.144739 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.183243 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcffbb96d-q8vwl" event={"ID":"e993934f-08d2-40c2-a9af-1b262dc941e9","Type":"ContainerStarted","Data":"ce527b9f4bc737de090171b379c1cdd131337ebf29ed93d441504407327cfdf8"} Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.184125 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.184406 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.188300 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-bcffbb96d-q8vwl" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.179:9311/healthcheck\": dial tcp 10.217.0.179:9311: connect: connection refused" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.193225 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" event={"ID":"7d1a6f4c-3bf3-4601-b2fe-62373ee28009","Type":"ContainerStarted","Data":"b66dcc924a532471b970cab2857d73bf3fe5cf105eecc0de903d55ef1973c637"} Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.197353 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d9b47bb8-6fqn7" event={"ID":"a5a47435-c445-47ff-8806-aa422090813c","Type":"ContainerStarted","Data":"fb23933ee3023e829dbcbf4c0a0602af977cacafc66c805de1fff16c6ca91111"} Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.216638 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" event={"ID":"eedf35c9-ac35-4a92-a1ea-c12beea2588b","Type":"ContainerStarted","Data":"3c775cacd95e4195e66cece96d3b1fa7c9f9599175f7bbefb9ac9b2a6fb371d8"} Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.225054 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-bcffbb96d-q8vwl" podStartSLOduration=7.225036082 podStartE2EDuration="7.225036082s" podCreationTimestamp="2025-12-03 17:36:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:36:32.205982546 +0000 UTC m=+1409.023453505" watchObservedRunningTime="2025-12-03 17:36:32.225036082 +0000 UTC m=+1409.042507041" Dec 03 17:36:32 crc kubenswrapper[4787]: I1203 17:36:32.696500 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-68677f5b6b-9rwg7"] Dec 03 17:36:33 crc kubenswrapper[4787]: I1203 17:36:33.227270 4787 generic.go:334] "Generic (PLEG): container finished" podID="eedf35c9-ac35-4a92-a1ea-c12beea2588b" containerID="3c775cacd95e4195e66cece96d3b1fa7c9f9599175f7bbefb9ac9b2a6fb371d8" exitCode=0 Dec 03 17:36:33 crc kubenswrapper[4787]: I1203 17:36:33.227364 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" event={"ID":"eedf35c9-ac35-4a92-a1ea-c12beea2588b","Type":"ContainerDied","Data":"3c775cacd95e4195e66cece96d3b1fa7c9f9599175f7bbefb9ac9b2a6fb371d8"} Dec 03 17:36:33 crc kubenswrapper[4787]: I1203 17:36:33.229684 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9fd5cd74f-xdkkz" event={"ID":"593e4809-69d6-43bf-ae94-a5d96431ca2f","Type":"ContainerStarted","Data":"fda77d07d025fb83299508a397fdaf7824bad2a3f1f4c016e44460c0771926de"} Dec 03 17:36:33 crc kubenswrapper[4787]: I1203 17:36:33.231217 4787 generic.go:334] "Generic (PLEG): container finished" podID="7d1a6f4c-3bf3-4601-b2fe-62373ee28009" containerID="b66dcc924a532471b970cab2857d73bf3fe5cf105eecc0de903d55ef1973c637" exitCode=0 Dec 03 17:36:33 crc kubenswrapper[4787]: I1203 17:36:33.231284 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" event={"ID":"7d1a6f4c-3bf3-4601-b2fe-62373ee28009","Type":"ContainerDied","Data":"b66dcc924a532471b970cab2857d73bf3fe5cf105eecc0de903d55ef1973c637"} Dec 03 17:36:33 crc kubenswrapper[4787]: I1203 17:36:33.245325 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68677f5b6b-9rwg7" event={"ID":"9dc4396d-2f35-42e0-a406-4f0a112fe1cd","Type":"ContainerStarted","Data":"3a3d9093012be8f6c7a73c3bd45e8628f9155c7caeb615cbb7fcaf6009ed9f12"} Dec 03 17:36:33 crc kubenswrapper[4787]: I1203 17:36:33.245383 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68677f5b6b-9rwg7" event={"ID":"9dc4396d-2f35-42e0-a406-4f0a112fe1cd","Type":"ContainerStarted","Data":"c1e815c7473362ce723d0d0159b1e1e2388046750f53913242ca16b3c68bdcbb"} Dec 03 17:36:33 crc kubenswrapper[4787]: I1203 17:36:33.257770 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d9b47bb8-6fqn7" event={"ID":"a5a47435-c445-47ff-8806-aa422090813c","Type":"ContainerStarted","Data":"6a933cc3673f14caef8f0be371decbf1bd9d8f5be6bcd9b1e80838a264078a2a"} Dec 03 17:36:33 crc kubenswrapper[4787]: I1203 17:36:33.950538 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.123774 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-sb\") pod \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.124589 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-swift-storage-0\") pod \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.124833 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wpg7\" (UniqueName: \"kubernetes.io/projected/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-kube-api-access-4wpg7\") pod \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.125096 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-svc\") pod \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.125297 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-config\") pod \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.125462 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-nb\") pod \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\" (UID: \"7d1a6f4c-3bf3-4601-b2fe-62373ee28009\") " Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.128907 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-kube-api-access-4wpg7" (OuterVolumeSpecName: "kube-api-access-4wpg7") pod "7d1a6f4c-3bf3-4601-b2fe-62373ee28009" (UID: "7d1a6f4c-3bf3-4601-b2fe-62373ee28009"). InnerVolumeSpecName "kube-api-access-4wpg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.151168 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-config" (OuterVolumeSpecName: "config") pod "7d1a6f4c-3bf3-4601-b2fe-62373ee28009" (UID: "7d1a6f4c-3bf3-4601-b2fe-62373ee28009"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.156100 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7d1a6f4c-3bf3-4601-b2fe-62373ee28009" (UID: "7d1a6f4c-3bf3-4601-b2fe-62373ee28009"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.164768 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7d1a6f4c-3bf3-4601-b2fe-62373ee28009" (UID: "7d1a6f4c-3bf3-4601-b2fe-62373ee28009"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.166135 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7d1a6f4c-3bf3-4601-b2fe-62373ee28009" (UID: "7d1a6f4c-3bf3-4601-b2fe-62373ee28009"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.169716 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7d1a6f4c-3bf3-4601-b2fe-62373ee28009" (UID: "7d1a6f4c-3bf3-4601-b2fe-62373ee28009"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.227858 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.227886 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.227897 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.227907 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.227917 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.227926 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wpg7\" (UniqueName: \"kubernetes.io/projected/7d1a6f4c-3bf3-4601-b2fe-62373ee28009-kube-api-access-4wpg7\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.304456 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.304580 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586bdc5f9-5cq5p" event={"ID":"7d1a6f4c-3bf3-4601-b2fe-62373ee28009","Type":"ContainerDied","Data":"8a96b5ae95157247a99881d5fa45722e6932547129a6a4a5214e0b3d8c673e7f"} Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.304655 4787 scope.go:117] "RemoveContainer" containerID="b66dcc924a532471b970cab2857d73bf3fe5cf105eecc0de903d55ef1973c637" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.304812 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.341432 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-69d9b47bb8-6fqn7" podStartSLOduration=9.341403561 podStartE2EDuration="9.341403561s" podCreationTimestamp="2025-12-03 17:36:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:36:34.333394588 +0000 UTC m=+1411.150865547" watchObservedRunningTime="2025-12-03 17:36:34.341403561 +0000 UTC m=+1411.158874520" Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.421348 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-5cq5p"] Dec 03 17:36:34 crc kubenswrapper[4787]: I1203 17:36:34.429914 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-586bdc5f9-5cq5p"] Dec 03 17:36:35 crc kubenswrapper[4787]: I1203 17:36:35.315900 4787 generic.go:334] "Generic (PLEG): container finished" podID="372bef16-5eab-41db-acbf-07274bf3f985" containerID="b9dfc7706b50afd1dd55f3a25a7644cf617bc170fecd695147bb42b76c618dc5" exitCode=0 Dec 03 17:36:35 crc kubenswrapper[4787]: I1203 17:36:35.315930 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fkrgl" event={"ID":"372bef16-5eab-41db-acbf-07274bf3f985","Type":"ContainerDied","Data":"b9dfc7706b50afd1dd55f3a25a7644cf617bc170fecd695147bb42b76c618dc5"} Dec 03 17:36:35 crc kubenswrapper[4787]: I1203 17:36:35.419689 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:35 crc kubenswrapper[4787]: I1203 17:36:35.782410 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d1a6f4c-3bf3-4601-b2fe-62373ee28009" path="/var/lib/kubelet/pods/7d1a6f4c-3bf3-4601-b2fe-62373ee28009/volumes" Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.730997 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fkrgl" Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.737434 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h596f\" (UniqueName: \"kubernetes.io/projected/372bef16-5eab-41db-acbf-07274bf3f985-kube-api-access-h596f\") pod \"372bef16-5eab-41db-acbf-07274bf3f985\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.737494 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-scripts\") pod \"372bef16-5eab-41db-acbf-07274bf3f985\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.737551 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-config-data\") pod \"372bef16-5eab-41db-acbf-07274bf3f985\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.737574 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-combined-ca-bundle\") pod \"372bef16-5eab-41db-acbf-07274bf3f985\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.737766 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372bef16-5eab-41db-acbf-07274bf3f985-logs\") pod \"372bef16-5eab-41db-acbf-07274bf3f985\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.738169 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/372bef16-5eab-41db-acbf-07274bf3f985-logs" (OuterVolumeSpecName: "logs") pod "372bef16-5eab-41db-acbf-07274bf3f985" (UID: "372bef16-5eab-41db-acbf-07274bf3f985"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.748088 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-scripts" (OuterVolumeSpecName: "scripts") pod "372bef16-5eab-41db-acbf-07274bf3f985" (UID: "372bef16-5eab-41db-acbf-07274bf3f985"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.754462 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/372bef16-5eab-41db-acbf-07274bf3f985-kube-api-access-h596f" (OuterVolumeSpecName: "kube-api-access-h596f") pod "372bef16-5eab-41db-acbf-07274bf3f985" (UID: "372bef16-5eab-41db-acbf-07274bf3f985"). InnerVolumeSpecName "kube-api-access-h596f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:39 crc kubenswrapper[4787]: E1203 17:36:39.772240 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-combined-ca-bundle podName:372bef16-5eab-41db-acbf-07274bf3f985 nodeName:}" failed. No retries permitted until 2025-12-03 17:36:40.272207347 +0000 UTC m=+1417.089678306 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-combined-ca-bundle") pod "372bef16-5eab-41db-acbf-07274bf3f985" (UID: "372bef16-5eab-41db-acbf-07274bf3f985") : error deleting /var/lib/kubelet/pods/372bef16-5eab-41db-acbf-07274bf3f985/volume-subpaths: remove /var/lib/kubelet/pods/372bef16-5eab-41db-acbf-07274bf3f985/volume-subpaths: no such file or directory Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.775260 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-config-data" (OuterVolumeSpecName: "config-data") pod "372bef16-5eab-41db-acbf-07274bf3f985" (UID: "372bef16-5eab-41db-acbf-07274bf3f985"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.840165 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h596f\" (UniqueName: \"kubernetes.io/projected/372bef16-5eab-41db-acbf-07274bf3f985-kube-api-access-h596f\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.840396 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.840405 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:39 crc kubenswrapper[4787]: I1203 17:36:39.840416 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/372bef16-5eab-41db-acbf-07274bf3f985-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.350004 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-combined-ca-bundle\") pod \"372bef16-5eab-41db-acbf-07274bf3f985\" (UID: \"372bef16-5eab-41db-acbf-07274bf3f985\") " Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.359817 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "372bef16-5eab-41db-acbf-07274bf3f985" (UID: "372bef16-5eab-41db-acbf-07274bf3f985"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.383600 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" event={"ID":"eedf35c9-ac35-4a92-a1ea-c12beea2588b","Type":"ContainerStarted","Data":"2ecbd44755df8e3613d7fcb80a72901141467122155521deb672f5b570d6a7ce"} Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.383979 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.395345 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-fkrgl" event={"ID":"372bef16-5eab-41db-acbf-07274bf3f985","Type":"ContainerDied","Data":"0cdf1e4bab7d994495e291f6954221ac6e48773ac0c0ed9b359b30af7114616c"} Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.395415 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cdf1e4bab7d994495e291f6954221ac6e48773ac0c0ed9b359b30af7114616c" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.395367 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-fkrgl" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.408597 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" podStartSLOduration=15.408580575 podStartE2EDuration="15.408580575s" podCreationTimestamp="2025-12-03 17:36:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:36:40.402718889 +0000 UTC m=+1417.220189838" watchObservedRunningTime="2025-12-03 17:36:40.408580575 +0000 UTC m=+1417.226051534" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.452885 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/372bef16-5eab-41db-acbf-07274bf3f985-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.710334 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.861111 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-65d88cc55b-crfnv"] Dec 03 17:36:40 crc kubenswrapper[4787]: E1203 17:36:40.861557 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d1a6f4c-3bf3-4601-b2fe-62373ee28009" containerName="init" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.861567 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d1a6f4c-3bf3-4601-b2fe-62373ee28009" containerName="init" Dec 03 17:36:40 crc kubenswrapper[4787]: E1203 17:36:40.861582 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="372bef16-5eab-41db-acbf-07274bf3f985" containerName="placement-db-sync" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.862843 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="372bef16-5eab-41db-acbf-07274bf3f985" containerName="placement-db-sync" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.863242 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d1a6f4c-3bf3-4601-b2fe-62373ee28009" containerName="init" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.863383 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="372bef16-5eab-41db-acbf-07274bf3f985" containerName="placement-db-sync" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.864947 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.869930 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-6nfqd" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.870284 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.870405 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.870507 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.870606 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 03 17:36:40 crc kubenswrapper[4787]: E1203 17:36:40.897928 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ceilometer-central-agent\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ceilometer-0" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" Dec 03 17:36:40 crc kubenswrapper[4787]: I1203 17:36:40.903653 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65d88cc55b-crfnv"] Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.072859 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-internal-tls-certs\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.073217 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea6b1667-020b-4078-894d-1b944b308802-logs\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.073282 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6bvq\" (UniqueName: \"kubernetes.io/projected/ea6b1667-020b-4078-894d-1b944b308802-kube-api-access-b6bvq\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.073317 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-combined-ca-bundle\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.073339 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-scripts\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.073371 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-config-data\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.073391 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-public-tls-certs\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.174816 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6bvq\" (UniqueName: \"kubernetes.io/projected/ea6b1667-020b-4078-894d-1b944b308802-kube-api-access-b6bvq\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.174890 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-combined-ca-bundle\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.174918 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-scripts\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.174957 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-config-data\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.174981 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-public-tls-certs\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.175094 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-internal-tls-certs\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.175159 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea6b1667-020b-4078-894d-1b944b308802-logs\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.184579 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-combined-ca-bundle\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.185399 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-scripts\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.186059 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-config-data\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.188255 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea6b1667-020b-4078-894d-1b944b308802-logs\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.189539 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-public-tls-certs\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.191268 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ea6b1667-020b-4078-894d-1b944b308802-internal-tls-certs\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.200732 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6bvq\" (UniqueName: \"kubernetes.io/projected/ea6b1667-020b-4078-894d-1b944b308802-kube-api-access-b6bvq\") pod \"placement-65d88cc55b-crfnv\" (UID: \"ea6b1667-020b-4078-894d-1b944b308802\") " pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.212214 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.426796 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" event={"ID":"d83ca165-acb3-4e27-b8f3-519897743134","Type":"ContainerStarted","Data":"6dcb7c4311d8146078eb401f50ffd3f291f67b82dfea434c90ed85a81355ef4e"} Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.443586 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-9fd5cd74f-xdkkz" event={"ID":"593e4809-69d6-43bf-ae94-a5d96431ca2f","Type":"ContainerStarted","Data":"5dc6914a9b30982481b3febc4a2a15e0c1f0c3058e6d9a5f92cc207b76ce2cca"} Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.517346 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="ceilometer-notification-agent" containerID="cri-o://08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38" gracePeriod=30 Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.518234 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="proxy-httpd" containerID="cri-o://25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651" gracePeriod=30 Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.518303 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="sg-core" containerID="cri-o://f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748" gracePeriod=30 Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.520756 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b18d92e-dfd0-4c8c-bb33-9af70f665417","Type":"ContainerStarted","Data":"25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651"} Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.520813 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.560929 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-9fd5cd74f-xdkkz" podStartSLOduration=13.56090707 podStartE2EDuration="13.56090707s" podCreationTimestamp="2025-12-03 17:36:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:36:41.487177331 +0000 UTC m=+1418.304648310" watchObservedRunningTime="2025-12-03 17:36:41.56090707 +0000 UTC m=+1418.378378029" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.563651 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5f6df7d97f-5vmdt" event={"ID":"65180fdc-f51b-45e7-ac70-05b9489e1201","Type":"ContainerStarted","Data":"540e8d47942875ece941710355e0bfa10ae5d3c906f440cc2703f4565d653c3f"} Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.563698 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5f6df7d97f-5vmdt" event={"ID":"65180fdc-f51b-45e7-ac70-05b9489e1201","Type":"ContainerStarted","Data":"ef88a5c7e6da49a925324acca8581e758e543a4ce4a2e552ae17b87ba4359b49"} Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.578129 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-68677f5b6b-9rwg7" event={"ID":"9dc4396d-2f35-42e0-a406-4f0a112fe1cd","Type":"ContainerStarted","Data":"344ce88d3d788b1f0c1b193ecf1c626d8171201821f755aae7e6f320cabf8f38"} Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.579141 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.583406 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.621365 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6pgw8" event={"ID":"5e5395cf-a94f-4125-8d92-54b7cb9da3be","Type":"ContainerStarted","Data":"df9dd7c608fc2760c54a5d62fabff29c223f72749d743297ee102cb769f0c531"} Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.628142 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5f6df7d97f-5vmdt" podStartSLOduration=4.044916384 podStartE2EDuration="17.628120676s" podCreationTimestamp="2025-12-03 17:36:24 +0000 UTC" firstStartedPulling="2025-12-03 17:36:26.594728175 +0000 UTC m=+1403.412199134" lastFinishedPulling="2025-12-03 17:36:40.177932467 +0000 UTC m=+1416.995403426" observedRunningTime="2025-12-03 17:36:41.585449632 +0000 UTC m=+1418.402920581" watchObservedRunningTime="2025-12-03 17:36:41.628120676 +0000 UTC m=+1418.445591635" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.649709 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-68677f5b6b-9rwg7" podStartSLOduration=10.649692329 podStartE2EDuration="10.649692329s" podCreationTimestamp="2025-12-03 17:36:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:36:41.630706185 +0000 UTC m=+1418.448177134" watchObservedRunningTime="2025-12-03 17:36:41.649692329 +0000 UTC m=+1418.467163278" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.662793 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-6pgw8" podStartSLOduration=3.365927729 podStartE2EDuration="52.662777447s" podCreationTimestamp="2025-12-03 17:35:49 +0000 UTC" firstStartedPulling="2025-12-03 17:35:51.033284832 +0000 UTC m=+1367.850755791" lastFinishedPulling="2025-12-03 17:36:40.33013455 +0000 UTC m=+1417.147605509" observedRunningTime="2025-12-03 17:36:41.65610413 +0000 UTC m=+1418.473575089" watchObservedRunningTime="2025-12-03 17:36:41.662777447 +0000 UTC m=+1418.480248406" Dec 03 17:36:41 crc kubenswrapper[4787]: I1203 17:36:41.790800 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-65d88cc55b-crfnv"] Dec 03 17:36:42 crc kubenswrapper[4787]: I1203 17:36:42.639350 4787 generic.go:334] "Generic (PLEG): container finished" podID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerID="25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651" exitCode=0 Dec 03 17:36:42 crc kubenswrapper[4787]: I1203 17:36:42.639562 4787 generic.go:334] "Generic (PLEG): container finished" podID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerID="f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748" exitCode=2 Dec 03 17:36:42 crc kubenswrapper[4787]: I1203 17:36:42.639599 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b18d92e-dfd0-4c8c-bb33-9af70f665417","Type":"ContainerDied","Data":"25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651"} Dec 03 17:36:42 crc kubenswrapper[4787]: I1203 17:36:42.639623 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b18d92e-dfd0-4c8c-bb33-9af70f665417","Type":"ContainerDied","Data":"f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748"} Dec 03 17:36:42 crc kubenswrapper[4787]: I1203 17:36:42.644172 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65d88cc55b-crfnv" event={"ID":"ea6b1667-020b-4078-894d-1b944b308802","Type":"ContainerStarted","Data":"ef24c8cf69be4c7421a3e7694ace996c30ca3d1bb414c06ddf3d0187275ea4cc"} Dec 03 17:36:42 crc kubenswrapper[4787]: I1203 17:36:42.644449 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65d88cc55b-crfnv" event={"ID":"ea6b1667-020b-4078-894d-1b944b308802","Type":"ContainerStarted","Data":"8b30d2c29b1f02e948b59ac07104e39217adad1d378c76ca743d1594e988cfac"} Dec 03 17:36:42 crc kubenswrapper[4787]: I1203 17:36:42.659391 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zmx58" event={"ID":"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90","Type":"ContainerStarted","Data":"0551230c6ebe89aad9bca04ef713444f4765ff4aae8f2b2e78bd5faef6d51db3"} Dec 03 17:36:42 crc kubenswrapper[4787]: I1203 17:36:42.662009 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" event={"ID":"d83ca165-acb3-4e27-b8f3-519897743134","Type":"ContainerStarted","Data":"9c2a03dfdd22951315286b543fbe97cd6cb9cfef4a75a1fe10083f1546f2f35b"} Dec 03 17:36:42 crc kubenswrapper[4787]: I1203 17:36:42.662641 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:42 crc kubenswrapper[4787]: I1203 17:36:42.716940 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-zmx58" podStartSLOduration=4.631347548 podStartE2EDuration="53.716917043s" podCreationTimestamp="2025-12-03 17:35:49 +0000 UTC" firstStartedPulling="2025-12-03 17:35:51.260407296 +0000 UTC m=+1368.077878255" lastFinishedPulling="2025-12-03 17:36:40.345976791 +0000 UTC m=+1417.163447750" observedRunningTime="2025-12-03 17:36:42.684555213 +0000 UTC m=+1419.502026182" watchObservedRunningTime="2025-12-03 17:36:42.716917043 +0000 UTC m=+1419.534388002" Dec 03 17:36:42 crc kubenswrapper[4787]: I1203 17:36:42.733763 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-78b7c4ccf4-bxfhc" podStartSLOduration=5.482447606 podStartE2EDuration="18.73373967s" podCreationTimestamp="2025-12-03 17:36:24 +0000 UTC" firstStartedPulling="2025-12-03 17:36:26.926235292 +0000 UTC m=+1403.743706251" lastFinishedPulling="2025-12-03 17:36:40.177527356 +0000 UTC m=+1416.994998315" observedRunningTime="2025-12-03 17:36:42.722883631 +0000 UTC m=+1419.540354590" watchObservedRunningTime="2025-12-03 17:36:42.73373967 +0000 UTC m=+1419.551210629" Dec 03 17:36:43 crc kubenswrapper[4787]: I1203 17:36:43.673511 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-65d88cc55b-crfnv" event={"ID":"ea6b1667-020b-4078-894d-1b944b308802","Type":"ContainerStarted","Data":"db5525686e9690a689a35025cd27a0b4b68cc92ecf1df81442186cd0087f5dcc"} Dec 03 17:36:43 crc kubenswrapper[4787]: I1203 17:36:43.673576 4787 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 03 17:36:43 crc kubenswrapper[4787]: I1203 17:36:43.705773 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-65d88cc55b-crfnv" podStartSLOduration=3.705747124 podStartE2EDuration="3.705747124s" podCreationTimestamp="2025-12-03 17:36:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:36:43.69692906 +0000 UTC m=+1420.514400049" watchObservedRunningTime="2025-12-03 17:36:43.705747124 +0000 UTC m=+1420.523218103" Dec 03 17:36:43 crc kubenswrapper[4787]: I1203 17:36:43.885509 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.522103 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.581512 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-run-httpd\") pod \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.581607 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-config-data\") pod \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.581732 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7p7j2\" (UniqueName: \"kubernetes.io/projected/5b18d92e-dfd0-4c8c-bb33-9af70f665417-kube-api-access-7p7j2\") pod \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.581830 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-scripts\") pod \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.581885 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-log-httpd\") pod \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.581908 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-sg-core-conf-yaml\") pod \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.581932 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-combined-ca-bundle\") pod \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\" (UID: \"5b18d92e-dfd0-4c8c-bb33-9af70f665417\") " Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.583343 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "5b18d92e-dfd0-4c8c-bb33-9af70f665417" (UID: "5b18d92e-dfd0-4c8c-bb33-9af70f665417"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.583305 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "5b18d92e-dfd0-4c8c-bb33-9af70f665417" (UID: "5b18d92e-dfd0-4c8c-bb33-9af70f665417"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.588610 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b18d92e-dfd0-4c8c-bb33-9af70f665417-kube-api-access-7p7j2" (OuterVolumeSpecName: "kube-api-access-7p7j2") pod "5b18d92e-dfd0-4c8c-bb33-9af70f665417" (UID: "5b18d92e-dfd0-4c8c-bb33-9af70f665417"). InnerVolumeSpecName "kube-api-access-7p7j2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.590114 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-scripts" (OuterVolumeSpecName: "scripts") pod "5b18d92e-dfd0-4c8c-bb33-9af70f665417" (UID: "5b18d92e-dfd0-4c8c-bb33-9af70f665417"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.639754 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "5b18d92e-dfd0-4c8c-bb33-9af70f665417" (UID: "5b18d92e-dfd0-4c8c-bb33-9af70f665417"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.662421 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5b18d92e-dfd0-4c8c-bb33-9af70f665417" (UID: "5b18d92e-dfd0-4c8c-bb33-9af70f665417"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.684986 4787 generic.go:334] "Generic (PLEG): container finished" podID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerID="08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38" exitCode=0 Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.685514 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.686104 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-config-data" (OuterVolumeSpecName: "config-data") pod "5b18d92e-dfd0-4c8c-bb33-9af70f665417" (UID: "5b18d92e-dfd0-4c8c-bb33-9af70f665417"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.686154 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b18d92e-dfd0-4c8c-bb33-9af70f665417","Type":"ContainerDied","Data":"08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38"} Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.686181 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"5b18d92e-dfd0-4c8c-bb33-9af70f665417","Type":"ContainerDied","Data":"e92f9ea2a393f913af1aaa874d1eff0a6e4dbf1da5675d47e970bf5c1d033484"} Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.686196 4787 scope.go:117] "RemoveContainer" containerID="25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.686349 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.686367 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.688403 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.688559 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7p7j2\" (UniqueName: \"kubernetes.io/projected/5b18d92e-dfd0-4c8c-bb33-9af70f665417-kube-api-access-7p7j2\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.688578 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.688588 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.688606 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.688616 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5b18d92e-dfd0-4c8c-bb33-9af70f665417-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.688625 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/5b18d92e-dfd0-4c8c-bb33-9af70f665417-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.706295 4787 scope.go:117] "RemoveContainer" containerID="f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.723101 4787 scope.go:117] "RemoveContainer" containerID="08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.746002 4787 scope.go:117] "RemoveContainer" containerID="25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651" Dec 03 17:36:44 crc kubenswrapper[4787]: E1203 17:36:44.746460 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651\": container with ID starting with 25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651 not found: ID does not exist" containerID="25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.746503 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651"} err="failed to get container status \"25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651\": rpc error: code = NotFound desc = could not find container \"25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651\": container with ID starting with 25fd27c58b4fc7e059af59a72b54653be72eff90339296bbd4d507263aa4e651 not found: ID does not exist" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.746528 4787 scope.go:117] "RemoveContainer" containerID="f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748" Dec 03 17:36:44 crc kubenswrapper[4787]: E1203 17:36:44.746917 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748\": container with ID starting with f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748 not found: ID does not exist" containerID="f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.746957 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748"} err="failed to get container status \"f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748\": rpc error: code = NotFound desc = could not find container \"f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748\": container with ID starting with f95bc7c0f06b72879da7b85221bc3605d5e9689999c5ee94c97a0bdbeaf53748 not found: ID does not exist" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.746983 4787 scope.go:117] "RemoveContainer" containerID="08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38" Dec 03 17:36:44 crc kubenswrapper[4787]: E1203 17:36:44.747584 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38\": container with ID starting with 08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38 not found: ID does not exist" containerID="08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38" Dec 03 17:36:44 crc kubenswrapper[4787]: I1203 17:36:44.747609 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38"} err="failed to get container status \"08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38\": rpc error: code = NotFound desc = could not find container \"08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38\": container with ID starting with 08f3a6bfce92cb1d02e2377e6b56a2e8f5f5b68b2f33e38de44ec48193f5cf38 not found: ID does not exist" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.080660 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.098932 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.124625 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:36:45 crc kubenswrapper[4787]: E1203 17:36:45.125253 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="sg-core" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.125277 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="sg-core" Dec 03 17:36:45 crc kubenswrapper[4787]: E1203 17:36:45.125337 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="ceilometer-notification-agent" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.125348 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="ceilometer-notification-agent" Dec 03 17:36:45 crc kubenswrapper[4787]: E1203 17:36:45.125363 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="proxy-httpd" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.125372 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="proxy-httpd" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.125638 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="ceilometer-notification-agent" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.125664 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="proxy-httpd" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.125691 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" containerName="sg-core" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.129532 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.132517 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.132758 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.155521 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.196161 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.196245 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-scripts\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.196270 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-config-data\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.196292 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.196382 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mmvqg\" (UniqueName: \"kubernetes.io/projected/fb3be536-b558-4a55-b1ce-f20c67948e10-kube-api-access-mmvqg\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.196453 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-log-httpd\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.196491 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-run-httpd\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.298447 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-run-httpd\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.298571 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.298659 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-config-data\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.298690 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.298719 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-scripts\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.298861 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mmvqg\" (UniqueName: \"kubernetes.io/projected/fb3be536-b558-4a55-b1ce-f20c67948e10-kube-api-access-mmvqg\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.298978 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-log-httpd\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.299820 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-log-httpd\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.303737 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-run-httpd\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.317035 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-config-data\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.326707 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-scripts\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.328347 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.333796 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mmvqg\" (UniqueName: \"kubernetes.io/projected/fb3be536-b558-4a55-b1ce-f20c67948e10-kube-api-access-mmvqg\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.334424 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.472119 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.778082 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b18d92e-dfd0-4c8c-bb33-9af70f665417" path="/var/lib/kubelet/pods/5b18d92e-dfd0-4c8c-bb33-9af70f665417/volumes" Dec 03 17:36:45 crc kubenswrapper[4787]: I1203 17:36:45.959343 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:36:45 crc kubenswrapper[4787]: W1203 17:36:45.960709 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb3be536_b558_4a55_b1ce_f20c67948e10.slice/crio-dd15234213bdddd73e5ed84c7af849bbb8f882c17d30e9d36a7ffd4db54cbc1f WatchSource:0}: Error finding container dd15234213bdddd73e5ed84c7af849bbb8f882c17d30e9d36a7ffd4db54cbc1f: Status 404 returned error can't find the container with id dd15234213bdddd73e5ed84c7af849bbb8f882c17d30e9d36a7ffd4db54cbc1f Dec 03 17:36:46 crc kubenswrapper[4787]: I1203 17:36:46.653150 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:36:46 crc kubenswrapper[4787]: I1203 17:36:46.764194 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb3be536-b558-4a55-b1ce-f20c67948e10","Type":"ContainerStarted","Data":"dd15234213bdddd73e5ed84c7af849bbb8f882c17d30e9d36a7ffd4db54cbc1f"} Dec 03 17:36:46 crc kubenswrapper[4787]: I1203 17:36:46.772982 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6xx4r"] Dec 03 17:36:46 crc kubenswrapper[4787]: I1203 17:36:46.773221 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" podUID="43a5db18-9a22-465e-bd9a-4d96572ba17b" containerName="dnsmasq-dns" containerID="cri-o://239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320" gracePeriod=10 Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.503720 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.676463 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-svc\") pod \"43a5db18-9a22-465e-bd9a-4d96572ba17b\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.676583 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-sb\") pod \"43a5db18-9a22-465e-bd9a-4d96572ba17b\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.676786 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-swift-storage-0\") pod \"43a5db18-9a22-465e-bd9a-4d96572ba17b\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.676876 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-config\") pod \"43a5db18-9a22-465e-bd9a-4d96572ba17b\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.677040 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxg98\" (UniqueName: \"kubernetes.io/projected/43a5db18-9a22-465e-bd9a-4d96572ba17b-kube-api-access-xxg98\") pod \"43a5db18-9a22-465e-bd9a-4d96572ba17b\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.677126 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-nb\") pod \"43a5db18-9a22-465e-bd9a-4d96572ba17b\" (UID: \"43a5db18-9a22-465e-bd9a-4d96572ba17b\") " Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.688438 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43a5db18-9a22-465e-bd9a-4d96572ba17b-kube-api-access-xxg98" (OuterVolumeSpecName: "kube-api-access-xxg98") pod "43a5db18-9a22-465e-bd9a-4d96572ba17b" (UID: "43a5db18-9a22-465e-bd9a-4d96572ba17b"). InnerVolumeSpecName "kube-api-access-xxg98". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.742904 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "43a5db18-9a22-465e-bd9a-4d96572ba17b" (UID: "43a5db18-9a22-465e-bd9a-4d96572ba17b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.748439 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "43a5db18-9a22-465e-bd9a-4d96572ba17b" (UID: "43a5db18-9a22-465e-bd9a-4d96572ba17b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.758565 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "43a5db18-9a22-465e-bd9a-4d96572ba17b" (UID: "43a5db18-9a22-465e-bd9a-4d96572ba17b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.759500 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "43a5db18-9a22-465e-bd9a-4d96572ba17b" (UID: "43a5db18-9a22-465e-bd9a-4d96572ba17b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.771414 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-config" (OuterVolumeSpecName: "config") pod "43a5db18-9a22-465e-bd9a-4d96572ba17b" (UID: "43a5db18-9a22-465e-bd9a-4d96572ba17b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.780738 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.780782 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.780800 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxg98\" (UniqueName: \"kubernetes.io/projected/43a5db18-9a22-465e-bd9a-4d96572ba17b-kube-api-access-xxg98\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.780813 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.780825 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.780835 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43a5db18-9a22-465e-bd9a-4d96572ba17b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.795145 4787 generic.go:334] "Generic (PLEG): container finished" podID="43a5db18-9a22-465e-bd9a-4d96572ba17b" containerID="239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320" exitCode=0 Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.795221 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" event={"ID":"43a5db18-9a22-465e-bd9a-4d96572ba17b","Type":"ContainerDied","Data":"239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320"} Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.795253 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" event={"ID":"43a5db18-9a22-465e-bd9a-4d96572ba17b","Type":"ContainerDied","Data":"48dee910c038cebfbb50fdef997fe49ff6ba65ba9b3c3a5cc9b8e57d76b4f43e"} Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.795278 4787 scope.go:117] "RemoveContainer" containerID="239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.795409 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-785d8bcb8c-6xx4r" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.805520 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb3be536-b558-4a55-b1ce-f20c67948e10","Type":"ContainerStarted","Data":"e863c048d1f5611690721248ddef6735980514cd16d0d6ae882ee94c2a0998de"} Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.805574 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb3be536-b558-4a55-b1ce-f20c67948e10","Type":"ContainerStarted","Data":"536728db935e62e50d81a2c5b98d7a76d3b39fe90f7536a4819eef30f73bfd0f"} Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.817857 4787 generic.go:334] "Generic (PLEG): container finished" podID="5e5395cf-a94f-4125-8d92-54b7cb9da3be" containerID="df9dd7c608fc2760c54a5d62fabff29c223f72749d743297ee102cb769f0c531" exitCode=0 Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.817911 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6pgw8" event={"ID":"5e5395cf-a94f-4125-8d92-54b7cb9da3be","Type":"ContainerDied","Data":"df9dd7c608fc2760c54a5d62fabff29c223f72749d743297ee102cb769f0c531"} Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.840225 4787 scope.go:117] "RemoveContainer" containerID="463d12055eb81946bfdf419f00afe7fb8856391fba24fde484aaedc7a73727a4" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.864552 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6xx4r"] Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.873293 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-785d8bcb8c-6xx4r"] Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.886095 4787 scope.go:117] "RemoveContainer" containerID="239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320" Dec 03 17:36:47 crc kubenswrapper[4787]: E1203 17:36:47.886736 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320\": container with ID starting with 239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320 not found: ID does not exist" containerID="239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.886782 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320"} err="failed to get container status \"239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320\": rpc error: code = NotFound desc = could not find container \"239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320\": container with ID starting with 239d9f89a923fec3fa2113419a34b669cbd55737b7b3b8c0bb0f85a9bc602320 not found: ID does not exist" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.886806 4787 scope.go:117] "RemoveContainer" containerID="463d12055eb81946bfdf419f00afe7fb8856391fba24fde484aaedc7a73727a4" Dec 03 17:36:47 crc kubenswrapper[4787]: E1203 17:36:47.891178 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"463d12055eb81946bfdf419f00afe7fb8856391fba24fde484aaedc7a73727a4\": container with ID starting with 463d12055eb81946bfdf419f00afe7fb8856391fba24fde484aaedc7a73727a4 not found: ID does not exist" containerID="463d12055eb81946bfdf419f00afe7fb8856391fba24fde484aaedc7a73727a4" Dec 03 17:36:47 crc kubenswrapper[4787]: I1203 17:36:47.891234 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"463d12055eb81946bfdf419f00afe7fb8856391fba24fde484aaedc7a73727a4"} err="failed to get container status \"463d12055eb81946bfdf419f00afe7fb8856391fba24fde484aaedc7a73727a4\": rpc error: code = NotFound desc = could not find container \"463d12055eb81946bfdf419f00afe7fb8856391fba24fde484aaedc7a73727a4\": container with ID starting with 463d12055eb81946bfdf419f00afe7fb8856391fba24fde484aaedc7a73727a4 not found: ID does not exist" Dec 03 17:36:48 crc kubenswrapper[4787]: I1203 17:36:48.829910 4787 generic.go:334] "Generic (PLEG): container finished" podID="cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" containerID="0551230c6ebe89aad9bca04ef713444f4765ff4aae8f2b2e78bd5faef6d51db3" exitCode=0 Dec 03 17:36:48 crc kubenswrapper[4787]: I1203 17:36:48.829965 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zmx58" event={"ID":"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90","Type":"ContainerDied","Data":"0551230c6ebe89aad9bca04ef713444f4765ff4aae8f2b2e78bd5faef6d51db3"} Dec 03 17:36:48 crc kubenswrapper[4787]: I1203 17:36:48.849578 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb3be536-b558-4a55-b1ce-f20c67948e10","Type":"ContainerStarted","Data":"c152187ae65339c6eb3bd61300787d5a2150b6dbc472213946c01ec22db54d15"} Dec 03 17:36:48 crc kubenswrapper[4787]: I1203 17:36:48.989828 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:36:48 crc kubenswrapper[4787]: I1203 17:36:48.989884 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.062833 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-68677f5b6b-9rwg7" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.155436 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-bcffbb96d-q8vwl"] Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.159836 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-bcffbb96d-q8vwl" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerName="barbican-api-log" containerID="cri-o://5fc3208f89bdfbfbd0258302400b81497e03a419922d16845c61a526e87a0aef" gracePeriod=30 Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.159907 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-bcffbb96d-q8vwl" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerName="barbican-api" containerID="cri-o://ce527b9f4bc737de090171b379c1cdd131337ebf29ed93d441504407327cfdf8" gracePeriod=30 Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.401321 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6pgw8" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.425459 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-combined-ca-bundle\") pod \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.425815 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-px6sz\" (UniqueName: \"kubernetes.io/projected/5e5395cf-a94f-4125-8d92-54b7cb9da3be-kube-api-access-px6sz\") pod \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.425946 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-config-data\") pod \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\" (UID: \"5e5395cf-a94f-4125-8d92-54b7cb9da3be\") " Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.448202 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e5395cf-a94f-4125-8d92-54b7cb9da3be-kube-api-access-px6sz" (OuterVolumeSpecName: "kube-api-access-px6sz") pod "5e5395cf-a94f-4125-8d92-54b7cb9da3be" (UID: "5e5395cf-a94f-4125-8d92-54b7cb9da3be"). InnerVolumeSpecName "kube-api-access-px6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.475826 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5e5395cf-a94f-4125-8d92-54b7cb9da3be" (UID: "5e5395cf-a94f-4125-8d92-54b7cb9da3be"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.528968 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.529032 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-px6sz\" (UniqueName: \"kubernetes.io/projected/5e5395cf-a94f-4125-8d92-54b7cb9da3be-kube-api-access-px6sz\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.536196 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-config-data" (OuterVolumeSpecName: "config-data") pod "5e5395cf-a94f-4125-8d92-54b7cb9da3be" (UID: "5e5395cf-a94f-4125-8d92-54b7cb9da3be"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.631265 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5e5395cf-a94f-4125-8d92-54b7cb9da3be-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.779783 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43a5db18-9a22-465e-bd9a-4d96572ba17b" path="/var/lib/kubelet/pods/43a5db18-9a22-465e-bd9a-4d96572ba17b/volumes" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.860605 4787 generic.go:334] "Generic (PLEG): container finished" podID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerID="5fc3208f89bdfbfbd0258302400b81497e03a419922d16845c61a526e87a0aef" exitCode=143 Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.860677 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcffbb96d-q8vwl" event={"ID":"e993934f-08d2-40c2-a9af-1b262dc941e9","Type":"ContainerDied","Data":"5fc3208f89bdfbfbd0258302400b81497e03a419922d16845c61a526e87a0aef"} Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.863146 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb3be536-b558-4a55-b1ce-f20c67948e10","Type":"ContainerStarted","Data":"ebbca1e662829b7c30611980184f208b1c1dac07f7e369d55f6016eab924f980"} Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.864474 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.869511 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-6pgw8" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.869501 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-6pgw8" event={"ID":"5e5395cf-a94f-4125-8d92-54b7cb9da3be","Type":"ContainerDied","Data":"b4c95243415c5cb41025f51a0fb32ff227cccba5e5c87a758b7f542accf4234c"} Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.869905 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b4c95243415c5cb41025f51a0fb32ff227cccba5e5c87a758b7f542accf4234c" Dec 03 17:36:49 crc kubenswrapper[4787]: I1203 17:36:49.904785 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.183108984 podStartE2EDuration="4.904748351s" podCreationTimestamp="2025-12-03 17:36:45 +0000 UTC" firstStartedPulling="2025-12-03 17:36:45.963625783 +0000 UTC m=+1422.781096742" lastFinishedPulling="2025-12-03 17:36:49.68526515 +0000 UTC m=+1426.502736109" observedRunningTime="2025-12-03 17:36:49.894102548 +0000 UTC m=+1426.711573537" watchObservedRunningTime="2025-12-03 17:36:49.904748351 +0000 UTC m=+1426.722219320" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.227548 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zmx58" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.342995 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwnrs\" (UniqueName: \"kubernetes.io/projected/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-kube-api-access-hwnrs\") pod \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.343107 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-scripts\") pod \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.343155 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-db-sync-config-data\") pod \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.343172 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-etc-machine-id\") pod \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.343212 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-combined-ca-bundle\") pod \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.343271 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-config-data\") pod \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\" (UID: \"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90\") " Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.345325 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" (UID: "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.355155 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" (UID: "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.355228 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-scripts" (OuterVolumeSpecName: "scripts") pod "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" (UID: "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.355797 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-kube-api-access-hwnrs" (OuterVolumeSpecName: "kube-api-access-hwnrs") pod "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" (UID: "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90"). InnerVolumeSpecName "kube-api-access-hwnrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.376270 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" (UID: "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.402476 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-config-data" (OuterVolumeSpecName: "config-data") pod "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" (UID: "cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.444940 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwnrs\" (UniqueName: \"kubernetes.io/projected/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-kube-api-access-hwnrs\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.445035 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.445061 4787 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.445070 4787 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.445077 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.445085 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.880806 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-zmx58" event={"ID":"cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90","Type":"ContainerDied","Data":"8eb389a049d093a2f6d5ddddd9607bfbf99b8323d31d79fbf9740d50db91f927"} Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.881161 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8eb389a049d093a2f6d5ddddd9607bfbf99b8323d31d79fbf9740d50db91f927" Dec 03 17:36:50 crc kubenswrapper[4787]: I1203 17:36:50.880869 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-zmx58" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.246547 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:36:51 crc kubenswrapper[4787]: E1203 17:36:51.246948 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e5395cf-a94f-4125-8d92-54b7cb9da3be" containerName="heat-db-sync" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.246964 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e5395cf-a94f-4125-8d92-54b7cb9da3be" containerName="heat-db-sync" Dec 03 17:36:51 crc kubenswrapper[4787]: E1203 17:36:51.246978 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a5db18-9a22-465e-bd9a-4d96572ba17b" containerName="dnsmasq-dns" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.246985 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a5db18-9a22-465e-bd9a-4d96572ba17b" containerName="dnsmasq-dns" Dec 03 17:36:51 crc kubenswrapper[4787]: E1203 17:36:51.246998 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" containerName="cinder-db-sync" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.247004 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" containerName="cinder-db-sync" Dec 03 17:36:51 crc kubenswrapper[4787]: E1203 17:36:51.247045 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43a5db18-9a22-465e-bd9a-4d96572ba17b" containerName="init" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.247051 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="43a5db18-9a22-465e-bd9a-4d96572ba17b" containerName="init" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.247227 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e5395cf-a94f-4125-8d92-54b7cb9da3be" containerName="heat-db-sync" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.247238 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="43a5db18-9a22-465e-bd9a-4d96572ba17b" containerName="dnsmasq-dns" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.247257 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" containerName="cinder-db-sync" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.254461 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.259200 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-d8gx4" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.259300 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.259581 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.265661 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.270157 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.296543 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6slhs"] Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.298611 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.347729 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6slhs"] Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.372686 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0081d61-8f59-44ab-9049-dc9e0d750b81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.372847 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.372903 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-scripts\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.372992 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.373137 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2zmv\" (UniqueName: \"kubernetes.io/projected/f0081d61-8f59-44ab-9049-dc9e0d750b81-kube-api-access-b2zmv\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.373177 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.453861 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.458913 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.461886 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.468184 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.474976 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0081d61-8f59-44ab-9049-dc9e0d750b81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.475032 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fdnm4\" (UniqueName: \"kubernetes.io/projected/88c9db22-be72-4ff3-a4bf-361d634922b0-kube-api-access-fdnm4\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.475075 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.475112 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.475133 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.475160 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.475182 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-config\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.475197 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-scripts\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.475238 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.475292 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.475316 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2zmv\" (UniqueName: \"kubernetes.io/projected/f0081d61-8f59-44ab-9049-dc9e0d750b81-kube-api-access-b2zmv\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.475337 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.476995 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0081d61-8f59-44ab-9049-dc9e0d750b81-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.479946 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.481430 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.485772 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-scripts\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.501828 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.510664 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2zmv\" (UniqueName: \"kubernetes.io/projected/f0081d61-8f59-44ab-9049-dc9e0d750b81-kube-api-access-b2zmv\") pod \"cinder-scheduler-0\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.576474 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fdnm4\" (UniqueName: \"kubernetes.io/projected/88c9db22-be72-4ff3-a4bf-361d634922b0-kube-api-access-fdnm4\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.576540 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h66vg\" (UniqueName: \"kubernetes.io/projected/d774a24e-5451-4976-a0d5-f915ebe863bf-kube-api-access-h66vg\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.576568 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.576602 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.576622 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.576652 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.576763 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-config\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.576878 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.576900 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.576964 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-scripts\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.577141 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d774a24e-5451-4976-a0d5-f915ebe863bf-logs\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.577312 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d774a24e-5451-4976-a0d5-f915ebe863bf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.577356 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data-custom\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.577569 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-svc\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.577592 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-swift-storage-0\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.577699 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-nb\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.578058 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-config\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.578274 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-sb\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.579982 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.610159 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fdnm4\" (UniqueName: \"kubernetes.io/projected/88c9db22-be72-4ff3-a4bf-361d634922b0-kube-api-access-fdnm4\") pod \"dnsmasq-dns-5c9776ccc5-6slhs\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.621560 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.679237 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d774a24e-5451-4976-a0d5-f915ebe863bf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.679287 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data-custom\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.679332 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d774a24e-5451-4976-a0d5-f915ebe863bf-etc-machine-id\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.679373 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h66vg\" (UniqueName: \"kubernetes.io/projected/d774a24e-5451-4976-a0d5-f915ebe863bf-kube-api-access-h66vg\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.679427 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.679506 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.679544 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-scripts\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.679612 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d774a24e-5451-4976-a0d5-f915ebe863bf-logs\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.680547 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d774a24e-5451-4976-a0d5-f915ebe863bf-logs\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.685914 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data-custom\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.686810 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-scripts\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.688674 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.689124 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.700710 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h66vg\" (UniqueName: \"kubernetes.io/projected/d774a24e-5451-4976-a0d5-f915ebe863bf-kube-api-access-h66vg\") pod \"cinder-api-0\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " pod="openstack/cinder-api-0" Dec 03 17:36:51 crc kubenswrapper[4787]: I1203 17:36:51.763433 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.109070 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.237847 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6slhs"] Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.355813 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.372284 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-bcffbb96d-q8vwl" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.179:9311/healthcheck\": read tcp 10.217.0.2:43070->10.217.0.179:9311: read: connection reset by peer" Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.372358 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-bcffbb96d-q8vwl" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.179:9311/healthcheck\": read tcp 10.217.0.2:43056->10.217.0.179:9311: read: connection reset by peer" Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.915141 4787 generic.go:334] "Generic (PLEG): container finished" podID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerID="ce527b9f4bc737de090171b379c1cdd131337ebf29ed93d441504407327cfdf8" exitCode=0 Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.915498 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcffbb96d-q8vwl" event={"ID":"e993934f-08d2-40c2-a9af-1b262dc941e9","Type":"ContainerDied","Data":"ce527b9f4bc737de090171b379c1cdd131337ebf29ed93d441504407327cfdf8"} Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.945381 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-l98hp"] Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.958714 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.959009 4787 generic.go:334] "Generic (PLEG): container finished" podID="88c9db22-be72-4ff3-a4bf-361d634922b0" containerID="84bd3b8ce3a556bd11579c6ad475079452d283de51169129f3b775d5cb11c38f" exitCode=0 Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.959194 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" event={"ID":"88c9db22-be72-4ff3-a4bf-361d634922b0","Type":"ContainerDied","Data":"84bd3b8ce3a556bd11579c6ad475079452d283de51169129f3b775d5cb11c38f"} Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.959227 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" event={"ID":"88c9db22-be72-4ff3-a4bf-361d634922b0","Type":"ContainerStarted","Data":"a4aa3563276ddeb0e749c90c679b01ee2cb91f766bd987d726b6497f16fb8b6b"} Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.975164 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l98hp"] Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.983828 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d774a24e-5451-4976-a0d5-f915ebe863bf","Type":"ContainerStarted","Data":"3c3d748c3f5516fbef788f28e5f8bf6b24546bb4b0f791edb2b95c14ebb3d30c"} Dec 03 17:36:52 crc kubenswrapper[4787]: I1203 17:36:52.993704 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f0081d61-8f59-44ab-9049-dc9e0d750b81","Type":"ContainerStarted","Data":"20285c2ab80b7c012c7ebe4bd3577e59c03ea40c7f4fae91f7ee1588d6e2c541"} Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.140415 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-utilities\") pod \"redhat-operators-l98hp\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.140787 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vmhrf\" (UniqueName: \"kubernetes.io/projected/1a553bba-383c-4507-bf61-43363d87959b-kube-api-access-vmhrf\") pod \"redhat-operators-l98hp\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.140892 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-catalog-content\") pod \"redhat-operators-l98hp\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.253496 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-utilities\") pod \"redhat-operators-l98hp\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.253592 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vmhrf\" (UniqueName: \"kubernetes.io/projected/1a553bba-383c-4507-bf61-43363d87959b-kube-api-access-vmhrf\") pod \"redhat-operators-l98hp\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.253693 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-catalog-content\") pod \"redhat-operators-l98hp\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.254370 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-catalog-content\") pod \"redhat-operators-l98hp\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.254660 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-utilities\") pod \"redhat-operators-l98hp\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.291273 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.303374 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vmhrf\" (UniqueName: \"kubernetes.io/projected/1a553bba-383c-4507-bf61-43363d87959b-kube-api-access-vmhrf\") pod \"redhat-operators-l98hp\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.461951 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data\") pod \"e993934f-08d2-40c2-a9af-1b262dc941e9\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.462083 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data-custom\") pod \"e993934f-08d2-40c2-a9af-1b262dc941e9\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.462116 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e993934f-08d2-40c2-a9af-1b262dc941e9-logs\") pod \"e993934f-08d2-40c2-a9af-1b262dc941e9\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.462144 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qlgp\" (UniqueName: \"kubernetes.io/projected/e993934f-08d2-40c2-a9af-1b262dc941e9-kube-api-access-4qlgp\") pod \"e993934f-08d2-40c2-a9af-1b262dc941e9\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.462167 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-combined-ca-bundle\") pod \"e993934f-08d2-40c2-a9af-1b262dc941e9\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.463587 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e993934f-08d2-40c2-a9af-1b262dc941e9-logs" (OuterVolumeSpecName: "logs") pod "e993934f-08d2-40c2-a9af-1b262dc941e9" (UID: "e993934f-08d2-40c2-a9af-1b262dc941e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.469096 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e993934f-08d2-40c2-a9af-1b262dc941e9-kube-api-access-4qlgp" (OuterVolumeSpecName: "kube-api-access-4qlgp") pod "e993934f-08d2-40c2-a9af-1b262dc941e9" (UID: "e993934f-08d2-40c2-a9af-1b262dc941e9"). InnerVolumeSpecName "kube-api-access-4qlgp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.472732 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e993934f-08d2-40c2-a9af-1b262dc941e9" (UID: "e993934f-08d2-40c2-a9af-1b262dc941e9"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.564447 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.564790 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e993934f-08d2-40c2-a9af-1b262dc941e9-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.564806 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qlgp\" (UniqueName: \"kubernetes.io/projected/e993934f-08d2-40c2-a9af-1b262dc941e9-kube-api-access-4qlgp\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.582483 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e993934f-08d2-40c2-a9af-1b262dc941e9" (UID: "e993934f-08d2-40c2-a9af-1b262dc941e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.582533 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.668180 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data" (OuterVolumeSpecName: "config-data") pod "e993934f-08d2-40c2-a9af-1b262dc941e9" (UID: "e993934f-08d2-40c2-a9af-1b262dc941e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.674852 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data\") pod \"e993934f-08d2-40c2-a9af-1b262dc941e9\" (UID: \"e993934f-08d2-40c2-a9af-1b262dc941e9\") " Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.675921 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:53 crc kubenswrapper[4787]: W1203 17:36:53.676049 4787 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/e993934f-08d2-40c2-a9af-1b262dc941e9/volumes/kubernetes.io~secret/config-data Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.676064 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data" (OuterVolumeSpecName: "config-data") pod "e993934f-08d2-40c2-a9af-1b262dc941e9" (UID: "e993934f-08d2-40c2-a9af-1b262dc941e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:36:53 crc kubenswrapper[4787]: I1203 17:36:53.778878 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e993934f-08d2-40c2-a9af-1b262dc941e9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.011204 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.065374 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-bcffbb96d-q8vwl" event={"ID":"e993934f-08d2-40c2-a9af-1b262dc941e9","Type":"ContainerDied","Data":"13f236f815e8d83bc381faeea8856c5a7decde53ba8887420c1ebc83bceaa7d5"} Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.065463 4787 scope.go:117] "RemoveContainer" containerID="ce527b9f4bc737de090171b379c1cdd131337ebf29ed93d441504407327cfdf8" Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.065656 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-bcffbb96d-q8vwl" Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.121118 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" event={"ID":"88c9db22-be72-4ff3-a4bf-361d634922b0","Type":"ContainerStarted","Data":"73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc"} Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.122282 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.123796 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-bcffbb96d-q8vwl"] Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.145372 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d774a24e-5451-4976-a0d5-f915ebe863bf","Type":"ContainerStarted","Data":"39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5"} Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.155270 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-bcffbb96d-q8vwl"] Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.168767 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" podStartSLOduration=3.168742144 podStartE2EDuration="3.168742144s" podCreationTimestamp="2025-12-03 17:36:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:36:54.151187879 +0000 UTC m=+1430.968658838" watchObservedRunningTime="2025-12-03 17:36:54.168742144 +0000 UTC m=+1430.986213103" Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.193703 4787 scope.go:117] "RemoveContainer" containerID="5fc3208f89bdfbfbd0258302400b81497e03a419922d16845c61a526e87a0aef" Dec 03 17:36:54 crc kubenswrapper[4787]: I1203 17:36:54.249894 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-l98hp"] Dec 03 17:36:55 crc kubenswrapper[4787]: I1203 17:36:55.160805 4787 generic.go:334] "Generic (PLEG): container finished" podID="1a553bba-383c-4507-bf61-43363d87959b" containerID="08fbfdb08e49786a03ebfb4dd9a96224df91ef9566fc1276333133242e4227bc" exitCode=0 Dec 03 17:36:55 crc kubenswrapper[4787]: I1203 17:36:55.161142 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l98hp" event={"ID":"1a553bba-383c-4507-bf61-43363d87959b","Type":"ContainerDied","Data":"08fbfdb08e49786a03ebfb4dd9a96224df91ef9566fc1276333133242e4227bc"} Dec 03 17:36:55 crc kubenswrapper[4787]: I1203 17:36:55.161176 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l98hp" event={"ID":"1a553bba-383c-4507-bf61-43363d87959b","Type":"ContainerStarted","Data":"312c96a6c67f650c683373dcadc929669d5e0248c57f8c644911056358593733"} Dec 03 17:36:55 crc kubenswrapper[4787]: I1203 17:36:55.166853 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d774a24e-5451-4976-a0d5-f915ebe863bf","Type":"ContainerStarted","Data":"e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183"} Dec 03 17:36:55 crc kubenswrapper[4787]: I1203 17:36:55.166982 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d774a24e-5451-4976-a0d5-f915ebe863bf" containerName="cinder-api-log" containerID="cri-o://39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5" gracePeriod=30 Dec 03 17:36:55 crc kubenswrapper[4787]: I1203 17:36:55.167067 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 17:36:55 crc kubenswrapper[4787]: I1203 17:36:55.167093 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="d774a24e-5451-4976-a0d5-f915ebe863bf" containerName="cinder-api" containerID="cri-o://e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183" gracePeriod=30 Dec 03 17:36:55 crc kubenswrapper[4787]: I1203 17:36:55.178986 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f0081d61-8f59-44ab-9049-dc9e0d750b81","Type":"ContainerStarted","Data":"faafe255dd6ff5cbd593371ceb92965c3f55cd883a468c17cb8c41cb6f654fbe"} Dec 03 17:36:55 crc kubenswrapper[4787]: I1203 17:36:55.212768 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.212753596 podStartE2EDuration="4.212753596s" podCreationTimestamp="2025-12-03 17:36:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:36:55.207367851 +0000 UTC m=+1432.024838810" watchObservedRunningTime="2025-12-03 17:36:55.212753596 +0000 UTC m=+1432.030224555" Dec 03 17:36:55 crc kubenswrapper[4787]: I1203 17:36:55.838751 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" path="/var/lib/kubelet/pods/e993934f-08d2-40c2-a9af-1b262dc941e9/volumes" Dec 03 17:36:56 crc kubenswrapper[4787]: I1203 17:36:56.198242 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f0081d61-8f59-44ab-9049-dc9e0d750b81","Type":"ContainerStarted","Data":"eb9bfc1818d2ad711fee698e2b6886aa9646eee1d65d7b85c61d705fa447e9ff"} Dec 03 17:36:56 crc kubenswrapper[4787]: I1203 17:36:56.210937 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l98hp" event={"ID":"1a553bba-383c-4507-bf61-43363d87959b","Type":"ContainerStarted","Data":"151d9d02ef4416fa639752d1217ec7ca9801ed76dad2150a9945371330ae8255"} Dec 03 17:36:56 crc kubenswrapper[4787]: I1203 17:36:56.220738 4787 generic.go:334] "Generic (PLEG): container finished" podID="d774a24e-5451-4976-a0d5-f915ebe863bf" containerID="39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5" exitCode=143 Dec 03 17:36:56 crc kubenswrapper[4787]: I1203 17:36:56.220816 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d774a24e-5451-4976-a0d5-f915ebe863bf","Type":"ContainerDied","Data":"39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5"} Dec 03 17:36:56 crc kubenswrapper[4787]: I1203 17:36:56.228242 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.279333299 podStartE2EDuration="5.228222226s" podCreationTimestamp="2025-12-03 17:36:51 +0000 UTC" firstStartedPulling="2025-12-03 17:36:52.109628623 +0000 UTC m=+1428.927099592" lastFinishedPulling="2025-12-03 17:36:53.05851756 +0000 UTC m=+1429.875988519" observedRunningTime="2025-12-03 17:36:56.225242525 +0000 UTC m=+1433.042713494" watchObservedRunningTime="2025-12-03 17:36:56.228222226 +0000 UTC m=+1433.045693185" Dec 03 17:36:56 crc kubenswrapper[4787]: I1203 17:36:56.582231 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 17:36:56 crc kubenswrapper[4787]: I1203 17:36:56.929483 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:36:57 crc kubenswrapper[4787]: I1203 17:36:57.189141 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-69bdc498c7-wqxdv" Dec 03 17:36:57 crc kubenswrapper[4787]: I1203 17:36:57.232975 4787 generic.go:334] "Generic (PLEG): container finished" podID="1a553bba-383c-4507-bf61-43363d87959b" containerID="151d9d02ef4416fa639752d1217ec7ca9801ed76dad2150a9945371330ae8255" exitCode=0 Dec 03 17:36:57 crc kubenswrapper[4787]: I1203 17:36:57.233781 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l98hp" event={"ID":"1a553bba-383c-4507-bf61-43363d87959b","Type":"ContainerDied","Data":"151d9d02ef4416fa639752d1217ec7ca9801ed76dad2150a9945371330ae8255"} Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.245430 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l98hp" event={"ID":"1a553bba-383c-4507-bf61-43363d87959b","Type":"ContainerStarted","Data":"49faf87cf76ca219b67501fd02a55ff8a95c2185cdff111b6d5ae9b9a8a8b9ab"} Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.265753 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-l98hp" podStartSLOduration=3.5804182989999997 podStartE2EDuration="6.265736512s" podCreationTimestamp="2025-12-03 17:36:52 +0000 UTC" firstStartedPulling="2025-12-03 17:36:55.16302873 +0000 UTC m=+1431.980499689" lastFinishedPulling="2025-12-03 17:36:57.848346953 +0000 UTC m=+1434.665817902" observedRunningTime="2025-12-03 17:36:58.264106268 +0000 UTC m=+1435.081577237" watchObservedRunningTime="2025-12-03 17:36:58.265736512 +0000 UTC m=+1435.083207471" Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.978076 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 03 17:36:58 crc kubenswrapper[4787]: E1203 17:36:58.978751 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerName="barbican-api" Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.978764 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerName="barbican-api" Dec 03 17:36:58 crc kubenswrapper[4787]: E1203 17:36:58.978786 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerName="barbican-api-log" Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.978793 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerName="barbican-api-log" Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.978973 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerName="barbican-api-log" Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.978997 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e993934f-08d2-40c2-a9af-1b262dc941e9" containerName="barbican-api" Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.979725 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.981849 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.982296 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.982344 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-mcq8s" Dec 03 17:36:58 crc kubenswrapper[4787]: I1203 17:36:58.989156 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.119728 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff54f756-3848-48ec-a235-d2814ff8d7f8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.119854 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcf2p\" (UniqueName: \"kubernetes.io/projected/ff54f756-3848-48ec-a235-d2814ff8d7f8-kube-api-access-gcf2p\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.119894 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ff54f756-3848-48ec-a235-d2814ff8d7f8-openstack-config-secret\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.119919 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ff54f756-3848-48ec-a235-d2814ff8d7f8-openstack-config\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.221610 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcf2p\" (UniqueName: \"kubernetes.io/projected/ff54f756-3848-48ec-a235-d2814ff8d7f8-kube-api-access-gcf2p\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.221697 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ff54f756-3848-48ec-a235-d2814ff8d7f8-openstack-config-secret\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.221729 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ff54f756-3848-48ec-a235-d2814ff8d7f8-openstack-config\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.221880 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff54f756-3848-48ec-a235-d2814ff8d7f8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.223193 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/ff54f756-3848-48ec-a235-d2814ff8d7f8-openstack-config\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.229261 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/ff54f756-3848-48ec-a235-d2814ff8d7f8-openstack-config-secret\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.235655 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ff54f756-3848-48ec-a235-d2814ff8d7f8-combined-ca-bundle\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.249147 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcf2p\" (UniqueName: \"kubernetes.io/projected/ff54f756-3848-48ec-a235-d2814ff8d7f8-kube-api-access-gcf2p\") pod \"openstackclient\" (UID: \"ff54f756-3848-48ec-a235-d2814ff8d7f8\") " pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.296531 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.334254 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-9fd5cd74f-xdkkz" Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.397773 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69d9b47bb8-6fqn7"] Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.398153 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69d9b47bb8-6fqn7" podUID="a5a47435-c445-47ff-8806-aa422090813c" containerName="neutron-api" containerID="cri-o://fb23933ee3023e829dbcbf4c0a0602af977cacafc66c805de1fff16c6ca91111" gracePeriod=30 Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.398669 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-69d9b47bb8-6fqn7" podUID="a5a47435-c445-47ff-8806-aa422090813c" containerName="neutron-httpd" containerID="cri-o://6a933cc3673f14caef8f0be371decbf1bd9d8f5be6bcd9b1e80838a264078a2a" gracePeriod=30 Dec 03 17:36:59 crc kubenswrapper[4787]: I1203 17:36:59.853376 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 03 17:36:59 crc kubenswrapper[4787]: W1203 17:36:59.856095 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff54f756_3848_48ec_a235_d2814ff8d7f8.slice/crio-164bcef5a422f1cc4f4baf8b90c4c2d31b8a930195e26d3d92c1885a4200e365 WatchSource:0}: Error finding container 164bcef5a422f1cc4f4baf8b90c4c2d31b8a930195e26d3d92c1885a4200e365: Status 404 returned error can't find the container with id 164bcef5a422f1cc4f4baf8b90c4c2d31b8a930195e26d3d92c1885a4200e365 Dec 03 17:37:00 crc kubenswrapper[4787]: I1203 17:37:00.268523 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ff54f756-3848-48ec-a235-d2814ff8d7f8","Type":"ContainerStarted","Data":"164bcef5a422f1cc4f4baf8b90c4c2d31b8a930195e26d3d92c1885a4200e365"} Dec 03 17:37:01 crc kubenswrapper[4787]: I1203 17:37:01.625298 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:37:01 crc kubenswrapper[4787]: I1203 17:37:01.708526 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-tw5cn"] Dec 03 17:37:01 crc kubenswrapper[4787]: I1203 17:37:01.710412 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" podUID="eedf35c9-ac35-4a92-a1ea-c12beea2588b" containerName="dnsmasq-dns" containerID="cri-o://2ecbd44755df8e3613d7fcb80a72901141467122155521deb672f5b570d6a7ce" gracePeriod=10 Dec 03 17:37:01 crc kubenswrapper[4787]: I1203 17:37:01.949548 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.005727 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.307702 4787 generic.go:334] "Generic (PLEG): container finished" podID="a5a47435-c445-47ff-8806-aa422090813c" containerID="6a933cc3673f14caef8f0be371decbf1bd9d8f5be6bcd9b1e80838a264078a2a" exitCode=0 Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.307851 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d9b47bb8-6fqn7" event={"ID":"a5a47435-c445-47ff-8806-aa422090813c","Type":"ContainerDied","Data":"6a933cc3673f14caef8f0be371decbf1bd9d8f5be6bcd9b1e80838a264078a2a"} Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.318348 4787 generic.go:334] "Generic (PLEG): container finished" podID="eedf35c9-ac35-4a92-a1ea-c12beea2588b" containerID="2ecbd44755df8e3613d7fcb80a72901141467122155521deb672f5b570d6a7ce" exitCode=0 Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.318591 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" event={"ID":"eedf35c9-ac35-4a92-a1ea-c12beea2588b","Type":"ContainerDied","Data":"2ecbd44755df8e3613d7fcb80a72901141467122155521deb672f5b570d6a7ce"} Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.318650 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f0081d61-8f59-44ab-9049-dc9e0d750b81" containerName="probe" containerID="cri-o://eb9bfc1818d2ad711fee698e2b6886aa9646eee1d65d7b85c61d705fa447e9ff" gracePeriod=30 Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.318731 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f0081d61-8f59-44ab-9049-dc9e0d750b81" containerName="cinder-scheduler" containerID="cri-o://faafe255dd6ff5cbd593371ceb92965c3f55cd883a468c17cb8c41cb6f654fbe" gracePeriod=30 Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.449883 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.596866 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-swift-storage-0\") pod \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.596941 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wflpf\" (UniqueName: \"kubernetes.io/projected/eedf35c9-ac35-4a92-a1ea-c12beea2588b-kube-api-access-wflpf\") pod \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.597034 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-nb\") pod \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.597073 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-config\") pod \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.597097 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-sb\") pod \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.597195 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-svc\") pod \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.605084 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eedf35c9-ac35-4a92-a1ea-c12beea2588b-kube-api-access-wflpf" (OuterVolumeSpecName: "kube-api-access-wflpf") pod "eedf35c9-ac35-4a92-a1ea-c12beea2588b" (UID: "eedf35c9-ac35-4a92-a1ea-c12beea2588b"). InnerVolumeSpecName "kube-api-access-wflpf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.684616 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-config" (OuterVolumeSpecName: "config") pod "eedf35c9-ac35-4a92-a1ea-c12beea2588b" (UID: "eedf35c9-ac35-4a92-a1ea-c12beea2588b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.685534 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "eedf35c9-ac35-4a92-a1ea-c12beea2588b" (UID: "eedf35c9-ac35-4a92-a1ea-c12beea2588b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.698677 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eedf35c9-ac35-4a92-a1ea-c12beea2588b" (UID: "eedf35c9-ac35-4a92-a1ea-c12beea2588b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.699092 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-svc\") pod \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\" (UID: \"eedf35c9-ac35-4a92-a1ea-c12beea2588b\") " Dec 03 17:37:02 crc kubenswrapper[4787]: W1203 17:37:02.699235 4787 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/eedf35c9-ac35-4a92-a1ea-c12beea2588b/volumes/kubernetes.io~configmap/dns-svc Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.699259 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "eedf35c9-ac35-4a92-a1ea-c12beea2588b" (UID: "eedf35c9-ac35-4a92-a1ea-c12beea2588b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.699801 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.699826 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wflpf\" (UniqueName: \"kubernetes.io/projected/eedf35c9-ac35-4a92-a1ea-c12beea2588b-kube-api-access-wflpf\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.699842 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.699853 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.701737 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "eedf35c9-ac35-4a92-a1ea-c12beea2588b" (UID: "eedf35c9-ac35-4a92-a1ea-c12beea2588b"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.735313 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "eedf35c9-ac35-4a92-a1ea-c12beea2588b" (UID: "eedf35c9-ac35-4a92-a1ea-c12beea2588b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.801264 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:02 crc kubenswrapper[4787]: I1203 17:37:02.801301 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/eedf35c9-ac35-4a92-a1ea-c12beea2588b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:03 crc kubenswrapper[4787]: I1203 17:37:03.330456 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" event={"ID":"eedf35c9-ac35-4a92-a1ea-c12beea2588b","Type":"ContainerDied","Data":"b493e649326147ad0157d611cd34d66154f43d03102d626bb20f6f6461191a88"} Dec 03 17:37:03 crc kubenswrapper[4787]: I1203 17:37:03.330546 4787 scope.go:117] "RemoveContainer" containerID="2ecbd44755df8e3613d7fcb80a72901141467122155521deb672f5b570d6a7ce" Dec 03 17:37:03 crc kubenswrapper[4787]: I1203 17:37:03.330475 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-85ff748b95-tw5cn" Dec 03 17:37:03 crc kubenswrapper[4787]: I1203 17:37:03.336520 4787 generic.go:334] "Generic (PLEG): container finished" podID="f0081d61-8f59-44ab-9049-dc9e0d750b81" containerID="eb9bfc1818d2ad711fee698e2b6886aa9646eee1d65d7b85c61d705fa447e9ff" exitCode=0 Dec 03 17:37:03 crc kubenswrapper[4787]: I1203 17:37:03.336556 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f0081d61-8f59-44ab-9049-dc9e0d750b81","Type":"ContainerDied","Data":"eb9bfc1818d2ad711fee698e2b6886aa9646eee1d65d7b85c61d705fa447e9ff"} Dec 03 17:37:03 crc kubenswrapper[4787]: I1203 17:37:03.370512 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-tw5cn"] Dec 03 17:37:03 crc kubenswrapper[4787]: I1203 17:37:03.371702 4787 scope.go:117] "RemoveContainer" containerID="3c775cacd95e4195e66cece96d3b1fa7c9f9599175f7bbefb9ac9b2a6fb371d8" Dec 03 17:37:03 crc kubenswrapper[4787]: I1203 17:37:03.381919 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-85ff748b95-tw5cn"] Dec 03 17:37:03 crc kubenswrapper[4787]: I1203 17:37:03.582837 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:37:03 crc kubenswrapper[4787]: I1203 17:37:03.583112 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:37:03 crc kubenswrapper[4787]: I1203 17:37:03.779178 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eedf35c9-ac35-4a92-a1ea-c12beea2588b" path="/var/lib/kubelet/pods/eedf35c9-ac35-4a92-a1ea-c12beea2588b/volumes" Dec 03 17:37:04 crc kubenswrapper[4787]: I1203 17:37:04.602346 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 17:37:04 crc kubenswrapper[4787]: I1203 17:37:04.639231 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-l98hp" podUID="1a553bba-383c-4507-bf61-43363d87959b" containerName="registry-server" probeResult="failure" output=< Dec 03 17:37:04 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 17:37:04 crc kubenswrapper[4787]: > Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.202636 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-8555854d45-f7p99"] Dec 03 17:37:06 crc kubenswrapper[4787]: E1203 17:37:06.203789 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eedf35c9-ac35-4a92-a1ea-c12beea2588b" containerName="dnsmasq-dns" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.203808 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eedf35c9-ac35-4a92-a1ea-c12beea2588b" containerName="dnsmasq-dns" Dec 03 17:37:06 crc kubenswrapper[4787]: E1203 17:37:06.203831 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eedf35c9-ac35-4a92-a1ea-c12beea2588b" containerName="init" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.203841 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="eedf35c9-ac35-4a92-a1ea-c12beea2588b" containerName="init" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.204122 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="eedf35c9-ac35-4a92-a1ea-c12beea2588b" containerName="dnsmasq-dns" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.205096 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.213824 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-engine-config-data" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.214072 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-heat-dockercfg-wn2sj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.214867 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-config-data" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.232334 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-8555854d45-f7p99"] Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.278258 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.278331 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-combined-ca-bundle\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.278667 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jqkpr\" (UniqueName: \"kubernetes.io/projected/d7cb720c-c038-4d14-9e53-293022882fcf-kube-api-access-jqkpr\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.278750 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data-custom\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.379942 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data-custom\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.380096 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.380133 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-combined-ca-bundle\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.380150 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jqkpr\" (UniqueName: \"kubernetes.io/projected/d7cb720c-c038-4d14-9e53-293022882fcf-kube-api-access-jqkpr\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.391271 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jlk65"] Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.395036 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-combined-ca-bundle\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.395672 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data-custom\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.401984 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.414325 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.415157 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jqkpr\" (UniqueName: \"kubernetes.io/projected/d7cb720c-c038-4d14-9e53-293022882fcf-kube-api-access-jqkpr\") pod \"heat-engine-8555854d45-f7p99\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.428685 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5bcbf94467-l2njc"] Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.439741 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.443165 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-cfnapi-config-data" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.446989 4787 generic.go:334] "Generic (PLEG): container finished" podID="f0081d61-8f59-44ab-9049-dc9e0d750b81" containerID="faafe255dd6ff5cbd593371ceb92965c3f55cd883a468c17cb8c41cb6f654fbe" exitCode=0 Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.450188 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f0081d61-8f59-44ab-9049-dc9e0d750b81","Type":"ContainerDied","Data":"faafe255dd6ff5cbd593371ceb92965c3f55cd883a468c17cb8c41cb6f654fbe"} Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.472948 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jlk65"] Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.486954 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.487049 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-config\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.487067 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.487110 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.487177 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.487247 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zp4hn\" (UniqueName: \"kubernetes.io/projected/b67a4816-85d5-404a-a22d-6cfbc455d286-kube-api-access-zp4hn\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.489123 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5bcbf94467-l2njc"] Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.521160 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-75d5f9cb88-48rbj"] Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.522838 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.525463 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"heat-api-config-data" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.542798 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-75d5f9cb88-48rbj"] Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.562484 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589163 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data-custom\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589211 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-combined-ca-bundle\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589242 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-config\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589263 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589284 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5qrs\" (UniqueName: \"kubernetes.io/projected/dc14f381-14f6-4719-a0a9-69bfc737fcb5-kube-api-access-l5qrs\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589334 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589352 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-combined-ca-bundle\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589376 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589408 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589465 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnjw7\" (UniqueName: \"kubernetes.io/projected/455e1ae0-75b9-4c80-9580-7b2ba4c72483-kube-api-access-vnjw7\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589488 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data-custom\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589513 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zp4hn\" (UniqueName: \"kubernetes.io/projected/b67a4816-85d5-404a-a22d-6cfbc455d286-kube-api-access-zp4hn\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589536 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.589564 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.590465 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-swift-storage-0\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.590474 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-config\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.590982 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-svc\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.592088 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-nb\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.594703 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-sb\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.608844 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zp4hn\" (UniqueName: \"kubernetes.io/projected/b67a4816-85d5-404a-a22d-6cfbc455d286-kube-api-access-zp4hn\") pod \"dnsmasq-dns-7756b9d78c-jlk65\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.691255 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.691327 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data-custom\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.691351 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-combined-ca-bundle\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.691374 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5qrs\" (UniqueName: \"kubernetes.io/projected/dc14f381-14f6-4719-a0a9-69bfc737fcb5-kube-api-access-l5qrs\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.691420 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-combined-ca-bundle\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.691447 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.691512 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnjw7\" (UniqueName: \"kubernetes.io/projected/455e1ae0-75b9-4c80-9580-7b2ba4c72483-kube-api-access-vnjw7\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.691536 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data-custom\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.696892 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data-custom\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.698337 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.700558 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-combined-ca-bundle\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.700794 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data-custom\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.701771 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.704618 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-combined-ca-bundle\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.720691 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnjw7\" (UniqueName: \"kubernetes.io/projected/455e1ae0-75b9-4c80-9580-7b2ba4c72483-kube-api-access-vnjw7\") pod \"heat-api-75d5f9cb88-48rbj\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.721835 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5qrs\" (UniqueName: \"kubernetes.io/projected/dc14f381-14f6-4719-a0a9-69bfc737fcb5-kube-api-access-l5qrs\") pod \"heat-cfnapi-5bcbf94467-l2njc\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.817084 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.828047 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:06 crc kubenswrapper[4787]: I1203 17:37:06.849910 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.007073 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.100796 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-combined-ca-bundle\") pod \"f0081d61-8f59-44ab-9049-dc9e0d750b81\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.101219 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0081d61-8f59-44ab-9049-dc9e0d750b81-etc-machine-id\") pod \"f0081d61-8f59-44ab-9049-dc9e0d750b81\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.101297 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data-custom\") pod \"f0081d61-8f59-44ab-9049-dc9e0d750b81\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.101357 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-scripts\") pod \"f0081d61-8f59-44ab-9049-dc9e0d750b81\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.101456 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b2zmv\" (UniqueName: \"kubernetes.io/projected/f0081d61-8f59-44ab-9049-dc9e0d750b81-kube-api-access-b2zmv\") pod \"f0081d61-8f59-44ab-9049-dc9e0d750b81\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.101488 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data\") pod \"f0081d61-8f59-44ab-9049-dc9e0d750b81\" (UID: \"f0081d61-8f59-44ab-9049-dc9e0d750b81\") " Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.116889 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f0081d61-8f59-44ab-9049-dc9e0d750b81-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f0081d61-8f59-44ab-9049-dc9e0d750b81" (UID: "f0081d61-8f59-44ab-9049-dc9e0d750b81"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.122233 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f0081d61-8f59-44ab-9049-dc9e0d750b81-kube-api-access-b2zmv" (OuterVolumeSpecName: "kube-api-access-b2zmv") pod "f0081d61-8f59-44ab-9049-dc9e0d750b81" (UID: "f0081d61-8f59-44ab-9049-dc9e0d750b81"). InnerVolumeSpecName "kube-api-access-b2zmv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.127704 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f0081d61-8f59-44ab-9049-dc9e0d750b81" (UID: "f0081d61-8f59-44ab-9049-dc9e0d750b81"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.152839 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-scripts" (OuterVolumeSpecName: "scripts") pod "f0081d61-8f59-44ab-9049-dc9e0d750b81" (UID: "f0081d61-8f59-44ab-9049-dc9e0d750b81"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.222256 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.222281 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.222290 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b2zmv\" (UniqueName: \"kubernetes.io/projected/f0081d61-8f59-44ab-9049-dc9e0d750b81-kube-api-access-b2zmv\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.222300 4787 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f0081d61-8f59-44ab-9049-dc9e0d750b81-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.246674 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-8555854d45-f7p99"] Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.285646 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f0081d61-8f59-44ab-9049-dc9e0d750b81" (UID: "f0081d61-8f59-44ab-9049-dc9e0d750b81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.328154 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.349837 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data" (OuterVolumeSpecName: "config-data") pod "f0081d61-8f59-44ab-9049-dc9e0d750b81" (UID: "f0081d61-8f59-44ab-9049-dc9e0d750b81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.441505 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f0081d61-8f59-44ab-9049-dc9e0d750b81-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.484891 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f0081d61-8f59-44ab-9049-dc9e0d750b81","Type":"ContainerDied","Data":"20285c2ab80b7c012c7ebe4bd3577e59c03ea40c7f4fae91f7ee1588d6e2c541"} Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.484948 4787 scope.go:117] "RemoveContainer" containerID="eb9bfc1818d2ad711fee698e2b6886aa9646eee1d65d7b85c61d705fa447e9ff" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.485116 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.494730 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-8555854d45-f7p99" event={"ID":"d7cb720c-c038-4d14-9e53-293022882fcf","Type":"ContainerStarted","Data":"3283fe2bd2aa7f36942c05618ef501b2d91323d5979fa56f988dc484f1459932"} Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.553080 4787 scope.go:117] "RemoveContainer" containerID="faafe255dd6ff5cbd593371ceb92965c3f55cd883a468c17cb8c41cb6f654fbe" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.557683 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jlk65"] Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.598652 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.623118 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.700830 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:37:07 crc kubenswrapper[4787]: E1203 17:37:07.702551 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0081d61-8f59-44ab-9049-dc9e0d750b81" containerName="cinder-scheduler" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.702589 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0081d61-8f59-44ab-9049-dc9e0d750b81" containerName="cinder-scheduler" Dec 03 17:37:07 crc kubenswrapper[4787]: E1203 17:37:07.702655 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f0081d61-8f59-44ab-9049-dc9e0d750b81" containerName="probe" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.702669 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f0081d61-8f59-44ab-9049-dc9e0d750b81" containerName="probe" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.719503 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0081d61-8f59-44ab-9049-dc9e0d750b81" containerName="probe" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.735704 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f0081d61-8f59-44ab-9049-dc9e0d750b81" containerName="cinder-scheduler" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.745171 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.745318 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.747465 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.758702 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-75d5f9cb88-48rbj"] Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.792371 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f0081d61-8f59-44ab-9049-dc9e0d750b81" path="/var/lib/kubelet/pods/f0081d61-8f59-44ab-9049-dc9e0d750b81/volumes" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.825581 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5bcbf94467-l2njc"] Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.862175 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-config-data\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.862249 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5dbx\" (UniqueName: \"kubernetes.io/projected/9cb81af5-f082-4509-b5a5-b0a49301b75e-kube-api-access-s5dbx\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.862291 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.862422 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9cb81af5-f082-4509-b5a5-b0a49301b75e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.862509 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-scripts\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.863713 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.965257 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9cb81af5-f082-4509-b5a5-b0a49301b75e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.965338 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-scripts\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.965370 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.965390 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-config-data\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.965427 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5dbx\" (UniqueName: \"kubernetes.io/projected/9cb81af5-f082-4509-b5a5-b0a49301b75e-kube-api-access-s5dbx\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.965456 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.968377 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/9cb81af5-f082-4509-b5a5-b0a49301b75e-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.971726 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-config-data\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.972843 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.972932 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-scripts\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.973094 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9cb81af5-f082-4509-b5a5-b0a49301b75e-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:07 crc kubenswrapper[4787]: I1203 17:37:07.989561 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5dbx\" (UniqueName: \"kubernetes.io/projected/9cb81af5-f082-4509-b5a5-b0a49301b75e-kube-api-access-s5dbx\") pod \"cinder-scheduler-0\" (UID: \"9cb81af5-f082-4509-b5a5-b0a49301b75e\") " pod="openstack/cinder-scheduler-0" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.114244 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.173976 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.174314 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="ceilometer-central-agent" containerID="cri-o://536728db935e62e50d81a2c5b98d7a76d3b39fe90f7536a4819eef30f73bfd0f" gracePeriod=30 Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.175141 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="proxy-httpd" containerID="cri-o://ebbca1e662829b7c30611980184f208b1c1dac07f7e369d55f6016eab924f980" gracePeriod=30 Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.175217 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="sg-core" containerID="cri-o://c152187ae65339c6eb3bd61300787d5a2150b6dbc472213946c01ec22db54d15" gracePeriod=30 Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.175265 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="ceilometer-notification-agent" containerID="cri-o://e863c048d1f5611690721248ddef6735980514cd16d0d6ae882ee94c2a0998de" gracePeriod=30 Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.180282 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.510990 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-8555854d45-f7p99" event={"ID":"d7cb720c-c038-4d14-9e53-293022882fcf","Type":"ContainerStarted","Data":"1ceb46625355ff189730f2313117ef55d982bb0f121844b31fa0043239dee766"} Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.512478 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.530147 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-75d5f9cb88-48rbj" event={"ID":"455e1ae0-75b9-4c80-9580-7b2ba4c72483","Type":"ContainerStarted","Data":"ff728de928f4390ce704b9690b0fc0885b42104a3518c2fcdfc2c8b6389a6254"} Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.538836 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" event={"ID":"b67a4816-85d5-404a-a22d-6cfbc455d286","Type":"ContainerStarted","Data":"dd3f44d97441b04422c9b10ecc784e281e8bde8e7b954f2d5274a7972ddc9ef6"} Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.563381 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-8555854d45-f7p99" podStartSLOduration=2.563356555 podStartE2EDuration="2.563356555s" podCreationTimestamp="2025-12-03 17:37:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:37:08.546336174 +0000 UTC m=+1445.363807143" watchObservedRunningTime="2025-12-03 17:37:08.563356555 +0000 UTC m=+1445.380827514" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.565246 4787 generic.go:334] "Generic (PLEG): container finished" podID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerID="c152187ae65339c6eb3bd61300787d5a2150b6dbc472213946c01ec22db54d15" exitCode=2 Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.565307 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb3be536-b558-4a55-b1ce-f20c67948e10","Type":"ContainerDied","Data":"c152187ae65339c6eb3bd61300787d5a2150b6dbc472213946c01ec22db54d15"} Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.761874 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-65b7898d75-fjwnb"] Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.763549 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.770466 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.771292 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.783958 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.789710 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65b7898d75-fjwnb"] Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.888547 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-public-tls-certs\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.888692 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-run-httpd\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.888718 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-config-data\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.888806 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-internal-tls-certs\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.888836 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-etc-swift\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.888886 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-log-httpd\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.888923 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-combined-ca-bundle\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.889011 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6sbm\" (UniqueName: \"kubernetes.io/projected/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-kube-api-access-p6sbm\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.990626 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-run-httpd\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.991168 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-config-data\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.991270 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-run-httpd\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.991475 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-internal-tls-certs\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.991600 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-etc-swift\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.991705 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-log-httpd\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.991791 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-combined-ca-bundle\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.991919 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6sbm\" (UniqueName: \"kubernetes.io/projected/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-kube-api-access-p6sbm\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.992082 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-public-tls-certs\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.992261 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-log-httpd\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:08 crc kubenswrapper[4787]: I1203 17:37:08.999836 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-public-tls-certs\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.000640 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-etc-swift\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.002279 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-config-data\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.002880 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-combined-ca-bundle\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.011670 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-internal-tls-certs\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.037775 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6sbm\" (UniqueName: \"kubernetes.io/projected/cf9c3742-1e29-4f87-bb53-9a60ab3c14ee-kube-api-access-p6sbm\") pod \"swift-proxy-65b7898d75-fjwnb\" (UID: \"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee\") " pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.136112 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.587235 4787 generic.go:334] "Generic (PLEG): container finished" podID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerID="ebbca1e662829b7c30611980184f208b1c1dac07f7e369d55f6016eab924f980" exitCode=0 Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.587604 4787 generic.go:334] "Generic (PLEG): container finished" podID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerID="536728db935e62e50d81a2c5b98d7a76d3b39fe90f7536a4819eef30f73bfd0f" exitCode=0 Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.587659 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb3be536-b558-4a55-b1ce-f20c67948e10","Type":"ContainerDied","Data":"ebbca1e662829b7c30611980184f208b1c1dac07f7e369d55f6016eab924f980"} Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.587694 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb3be536-b558-4a55-b1ce-f20c67948e10","Type":"ContainerDied","Data":"536728db935e62e50d81a2c5b98d7a76d3b39fe90f7536a4819eef30f73bfd0f"} Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.594623 4787 generic.go:334] "Generic (PLEG): container finished" podID="a5a47435-c445-47ff-8806-aa422090813c" containerID="fb23933ee3023e829dbcbf4c0a0602af977cacafc66c805de1fff16c6ca91111" exitCode=0 Dec 03 17:37:09 crc kubenswrapper[4787]: I1203 17:37:09.595596 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d9b47bb8-6fqn7" event={"ID":"a5a47435-c445-47ff-8806-aa422090813c","Type":"ContainerDied","Data":"fb23933ee3023e829dbcbf4c0a0602af977cacafc66c805de1fff16c6ca91111"} Dec 03 17:37:11 crc kubenswrapper[4787]: I1203 17:37:11.631428 4787 generic.go:334] "Generic (PLEG): container finished" podID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerID="e863c048d1f5611690721248ddef6735980514cd16d0d6ae882ee94c2a0998de" exitCode=0 Dec 03 17:37:11 crc kubenswrapper[4787]: I1203 17:37:11.631602 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb3be536-b558-4a55-b1ce-f20c67948e10","Type":"ContainerDied","Data":"e863c048d1f5611690721248ddef6735980514cd16d0d6ae882ee94c2a0998de"} Dec 03 17:37:13 crc kubenswrapper[4787]: I1203 17:37:13.645387 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:37:13 crc kubenswrapper[4787]: I1203 17:37:13.696955 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:37:13 crc kubenswrapper[4787]: I1203 17:37:13.910521 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l98hp"] Dec 03 17:37:14 crc kubenswrapper[4787]: I1203 17:37:14.377187 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:37:14 crc kubenswrapper[4787]: I1203 17:37:14.440319 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-65d88cc55b-crfnv" Dec 03 17:37:14 crc kubenswrapper[4787]: I1203 17:37:14.673943 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-l98hp" podUID="1a553bba-383c-4507-bf61-43363d87959b" containerName="registry-server" containerID="cri-o://49faf87cf76ca219b67501fd02a55ff8a95c2185cdff111b6d5ae9b9a8a8b9ab" gracePeriod=2 Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.058876 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-8b787c4b7-wwr58"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.061047 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.093638 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-8b787c4b7-wwr58"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.118180 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-7994c4bbc4-2kvbc"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.119648 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.127907 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-64cb8b689b-d4pmz"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.139434 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.142489 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7994c4bbc4-2kvbc"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.155008 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-64cb8b689b-d4pmz"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.226222 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p27gf\" (UniqueName: \"kubernetes.io/projected/9296f967-7bf3-4e42-a867-cce34092dc11-kube-api-access-p27gf\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.226330 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.226353 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-combined-ca-bundle\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.227870 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data-custom\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.227967 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.227987 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data-custom\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.228058 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlkpj\" (UniqueName: \"kubernetes.io/projected/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-kube-api-access-xlkpj\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.228131 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-combined-ca-bundle\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.329731 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlkpj\" (UniqueName: \"kubernetes.io/projected/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-kube-api-access-xlkpj\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.329803 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-combined-ca-bundle\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.329861 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p27gf\" (UniqueName: \"kubernetes.io/projected/9296f967-7bf3-4e42-a867-cce34092dc11-kube-api-access-p27gf\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.329906 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.329937 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-combined-ca-bundle\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.329958 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7z7v\" (UniqueName: \"kubernetes.io/projected/8f12c699-048b-4eee-bc7c-e3ad916656ed-kube-api-access-d7z7v\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.329986 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-combined-ca-bundle\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.330005 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.330059 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data-custom\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.330097 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data-custom\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.330126 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.330142 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data-custom\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.337140 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-combined-ca-bundle\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.338161 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data-custom\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.340833 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data-custom\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.341379 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.341896 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.348809 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-combined-ca-bundle\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.349058 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlkpj\" (UniqueName: \"kubernetes.io/projected/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-kube-api-access-xlkpj\") pod \"heat-engine-8b787c4b7-wwr58\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.351413 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p27gf\" (UniqueName: \"kubernetes.io/projected/9296f967-7bf3-4e42-a867-cce34092dc11-kube-api-access-p27gf\") pod \"heat-api-7994c4bbc4-2kvbc\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.432039 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-combined-ca-bundle\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.432096 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7z7v\" (UniqueName: \"kubernetes.io/projected/8f12c699-048b-4eee-bc7c-e3ad916656ed-kube-api-access-d7z7v\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.432184 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data-custom\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.432341 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.434359 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-lcm5g"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.436085 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-lcm5g" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.437204 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data-custom\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.437474 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-combined-ca-bundle\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.439454 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.442463 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.453888 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-hj85k"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.455590 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hj85k" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.460007 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7z7v\" (UniqueName: \"kubernetes.io/projected/8f12c699-048b-4eee-bc7c-e3ad916656ed-kube-api-access-d7z7v\") pod \"heat-cfnapi-64cb8b689b-d4pmz\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.462175 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.477362 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.477836 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="proxy-httpd" probeResult="failure" output="Get \"http://10.217.0.185:3000/\": dial tcp 10.217.0.185:3000: connect: connection refused" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.478120 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hj85k"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.492088 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-lcm5g"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.554652 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-6dhs9"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.556940 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6dhs9" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.567929 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6dhs9"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.639151 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97mgf\" (UniqueName: \"kubernetes.io/projected/6371d53b-b035-4a6e-82b5-c5774ffc8de1-kube-api-access-97mgf\") pod \"nova-api-db-create-hj85k\" (UID: \"6371d53b-b035-4a6e-82b5-c5774ffc8de1\") " pod="openstack/nova-api-db-create-hj85k" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.639598 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b9s9\" (UniqueName: \"kubernetes.io/projected/244e40a3-b432-4971-bdf5-f070ab9891b0-kube-api-access-5b9s9\") pod \"nova-cell0-db-create-lcm5g\" (UID: \"244e40a3-b432-4971-bdf5-f070ab9891b0\") " pod="openstack/nova-cell0-db-create-lcm5g" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.639673 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/244e40a3-b432-4971-bdf5-f070ab9891b0-operator-scripts\") pod \"nova-cell0-db-create-lcm5g\" (UID: \"244e40a3-b432-4971-bdf5-f070ab9891b0\") " pod="openstack/nova-cell0-db-create-lcm5g" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.639885 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6371d53b-b035-4a6e-82b5-c5774ffc8de1-operator-scripts\") pod \"nova-api-db-create-hj85k\" (UID: \"6371d53b-b035-4a6e-82b5-c5774ffc8de1\") " pod="openstack/nova-api-db-create-hj85k" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.650967 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-3101-account-create-update-g8j4t"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.653661 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3101-account-create-update-g8j4t" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.661477 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.665352 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3101-account-create-update-g8j4t"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.705166 4787 generic.go:334] "Generic (PLEG): container finished" podID="1a553bba-383c-4507-bf61-43363d87959b" containerID="49faf87cf76ca219b67501fd02a55ff8a95c2185cdff111b6d5ae9b9a8a8b9ab" exitCode=0 Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.705211 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l98hp" event={"ID":"1a553bba-383c-4507-bf61-43363d87959b","Type":"ContainerDied","Data":"49faf87cf76ca219b67501fd02a55ff8a95c2185cdff111b6d5ae9b9a8a8b9ab"} Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.741307 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6371d53b-b035-4a6e-82b5-c5774ffc8de1-operator-scripts\") pod \"nova-api-db-create-hj85k\" (UID: \"6371d53b-b035-4a6e-82b5-c5774ffc8de1\") " pod="openstack/nova-api-db-create-hj85k" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.741358 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97mgf\" (UniqueName: \"kubernetes.io/projected/6371d53b-b035-4a6e-82b5-c5774ffc8de1-kube-api-access-97mgf\") pod \"nova-api-db-create-hj85k\" (UID: \"6371d53b-b035-4a6e-82b5-c5774ffc8de1\") " pod="openstack/nova-api-db-create-hj85k" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.741396 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a23593c-7762-4639-8b2f-2cdd49b0c728-operator-scripts\") pod \"nova-cell1-db-create-6dhs9\" (UID: \"1a23593c-7762-4639-8b2f-2cdd49b0c728\") " pod="openstack/nova-cell1-db-create-6dhs9" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.741432 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b9s9\" (UniqueName: \"kubernetes.io/projected/244e40a3-b432-4971-bdf5-f070ab9891b0-kube-api-access-5b9s9\") pod \"nova-cell0-db-create-lcm5g\" (UID: \"244e40a3-b432-4971-bdf5-f070ab9891b0\") " pod="openstack/nova-cell0-db-create-lcm5g" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.741472 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/244e40a3-b432-4971-bdf5-f070ab9891b0-operator-scripts\") pod \"nova-cell0-db-create-lcm5g\" (UID: \"244e40a3-b432-4971-bdf5-f070ab9891b0\") " pod="openstack/nova-cell0-db-create-lcm5g" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.741491 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bprnc\" (UniqueName: \"kubernetes.io/projected/1a23593c-7762-4639-8b2f-2cdd49b0c728-kube-api-access-bprnc\") pod \"nova-cell1-db-create-6dhs9\" (UID: \"1a23593c-7762-4639-8b2f-2cdd49b0c728\") " pod="openstack/nova-cell1-db-create-6dhs9" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.754800 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/244e40a3-b432-4971-bdf5-f070ab9891b0-operator-scripts\") pod \"nova-cell0-db-create-lcm5g\" (UID: \"244e40a3-b432-4971-bdf5-f070ab9891b0\") " pod="openstack/nova-cell0-db-create-lcm5g" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.755163 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6371d53b-b035-4a6e-82b5-c5774ffc8de1-operator-scripts\") pod \"nova-api-db-create-hj85k\" (UID: \"6371d53b-b035-4a6e-82b5-c5774ffc8de1\") " pod="openstack/nova-api-db-create-hj85k" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.763887 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b9s9\" (UniqueName: \"kubernetes.io/projected/244e40a3-b432-4971-bdf5-f070ab9891b0-kube-api-access-5b9s9\") pod \"nova-cell0-db-create-lcm5g\" (UID: \"244e40a3-b432-4971-bdf5-f070ab9891b0\") " pod="openstack/nova-cell0-db-create-lcm5g" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.779848 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97mgf\" (UniqueName: \"kubernetes.io/projected/6371d53b-b035-4a6e-82b5-c5774ffc8de1-kube-api-access-97mgf\") pod \"nova-api-db-create-hj85k\" (UID: \"6371d53b-b035-4a6e-82b5-c5774ffc8de1\") " pod="openstack/nova-api-db-create-hj85k" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.799197 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-a7a8-account-create-update-72nlv"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.800874 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.813606 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.817640 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a7a8-account-create-update-72nlv"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.843142 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbvk7\" (UniqueName: \"kubernetes.io/projected/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-kube-api-access-dbvk7\") pod \"nova-api-3101-account-create-update-g8j4t\" (UID: \"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a\") " pod="openstack/nova-api-3101-account-create-update-g8j4t" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.843192 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bprnc\" (UniqueName: \"kubernetes.io/projected/1a23593c-7762-4639-8b2f-2cdd49b0c728-kube-api-access-bprnc\") pod \"nova-cell1-db-create-6dhs9\" (UID: \"1a23593c-7762-4639-8b2f-2cdd49b0c728\") " pod="openstack/nova-cell1-db-create-6dhs9" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.843343 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-operator-scripts\") pod \"nova-api-3101-account-create-update-g8j4t\" (UID: \"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a\") " pod="openstack/nova-api-3101-account-create-update-g8j4t" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.843463 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a23593c-7762-4639-8b2f-2cdd49b0c728-operator-scripts\") pod \"nova-cell1-db-create-6dhs9\" (UID: \"1a23593c-7762-4639-8b2f-2cdd49b0c728\") " pod="openstack/nova-cell1-db-create-6dhs9" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.844399 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a23593c-7762-4639-8b2f-2cdd49b0c728-operator-scripts\") pod \"nova-cell1-db-create-6dhs9\" (UID: \"1a23593c-7762-4639-8b2f-2cdd49b0c728\") " pod="openstack/nova-cell1-db-create-6dhs9" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.869061 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bprnc\" (UniqueName: \"kubernetes.io/projected/1a23593c-7762-4639-8b2f-2cdd49b0c728-kube-api-access-bprnc\") pod \"nova-cell1-db-create-6dhs9\" (UID: \"1a23593c-7762-4639-8b2f-2cdd49b0c728\") " pod="openstack/nova-cell1-db-create-6dhs9" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.874918 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-lcm5g" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.898930 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hj85k" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.952296 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-91eb-account-create-update-59l2z"] Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.953839 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-91eb-account-create-update-59l2z" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.956434 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.960549 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ztbw9\" (UniqueName: \"kubernetes.io/projected/39227401-0c97-4fd7-8510-615e22ca73d9-kube-api-access-ztbw9\") pod \"nova-cell0-a7a8-account-create-update-72nlv\" (UID: \"39227401-0c97-4fd7-8510-615e22ca73d9\") " pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.960697 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39227401-0c97-4fd7-8510-615e22ca73d9-operator-scripts\") pod \"nova-cell0-a7a8-account-create-update-72nlv\" (UID: \"39227401-0c97-4fd7-8510-615e22ca73d9\") " pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.960764 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-operator-scripts\") pod \"nova-api-3101-account-create-update-g8j4t\" (UID: \"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a\") " pod="openstack/nova-api-3101-account-create-update-g8j4t" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.964594 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6dhs9" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.965483 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbvk7\" (UniqueName: \"kubernetes.io/projected/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-kube-api-access-dbvk7\") pod \"nova-api-3101-account-create-update-g8j4t\" (UID: \"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a\") " pod="openstack/nova-api-3101-account-create-update-g8j4t" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.967919 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-operator-scripts\") pod \"nova-api-3101-account-create-update-g8j4t\" (UID: \"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a\") " pod="openstack/nova-api-3101-account-create-update-g8j4t" Dec 03 17:37:15 crc kubenswrapper[4787]: I1203 17:37:15.985896 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-91eb-account-create-update-59l2z"] Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.010274 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbvk7\" (UniqueName: \"kubernetes.io/projected/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-kube-api-access-dbvk7\") pod \"nova-api-3101-account-create-update-g8j4t\" (UID: \"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a\") " pod="openstack/nova-api-3101-account-create-update-g8j4t" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.069681 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75bff842-ba6b-4f84-bff6-054b7292d82e-operator-scripts\") pod \"nova-cell1-91eb-account-create-update-59l2z\" (UID: \"75bff842-ba6b-4f84-bff6-054b7292d82e\") " pod="openstack/nova-cell1-91eb-account-create-update-59l2z" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.069754 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39227401-0c97-4fd7-8510-615e22ca73d9-operator-scripts\") pod \"nova-cell0-a7a8-account-create-update-72nlv\" (UID: \"39227401-0c97-4fd7-8510-615e22ca73d9\") " pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.069782 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nng29\" (UniqueName: \"kubernetes.io/projected/75bff842-ba6b-4f84-bff6-054b7292d82e-kube-api-access-nng29\") pod \"nova-cell1-91eb-account-create-update-59l2z\" (UID: \"75bff842-ba6b-4f84-bff6-054b7292d82e\") " pod="openstack/nova-cell1-91eb-account-create-update-59l2z" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.069886 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ztbw9\" (UniqueName: \"kubernetes.io/projected/39227401-0c97-4fd7-8510-615e22ca73d9-kube-api-access-ztbw9\") pod \"nova-cell0-a7a8-account-create-update-72nlv\" (UID: \"39227401-0c97-4fd7-8510-615e22ca73d9\") " pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.070887 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39227401-0c97-4fd7-8510-615e22ca73d9-operator-scripts\") pod \"nova-cell0-a7a8-account-create-update-72nlv\" (UID: \"39227401-0c97-4fd7-8510-615e22ca73d9\") " pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.101856 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ztbw9\" (UniqueName: \"kubernetes.io/projected/39227401-0c97-4fd7-8510-615e22ca73d9-kube-api-access-ztbw9\") pod \"nova-cell0-a7a8-account-create-update-72nlv\" (UID: \"39227401-0c97-4fd7-8510-615e22ca73d9\") " pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.155614 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.178364 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75bff842-ba6b-4f84-bff6-054b7292d82e-operator-scripts\") pod \"nova-cell1-91eb-account-create-update-59l2z\" (UID: \"75bff842-ba6b-4f84-bff6-054b7292d82e\") " pod="openstack/nova-cell1-91eb-account-create-update-59l2z" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.179198 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75bff842-ba6b-4f84-bff6-054b7292d82e-operator-scripts\") pod \"nova-cell1-91eb-account-create-update-59l2z\" (UID: \"75bff842-ba6b-4f84-bff6-054b7292d82e\") " pod="openstack/nova-cell1-91eb-account-create-update-59l2z" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.188352 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nng29\" (UniqueName: \"kubernetes.io/projected/75bff842-ba6b-4f84-bff6-054b7292d82e-kube-api-access-nng29\") pod \"nova-cell1-91eb-account-create-update-59l2z\" (UID: \"75bff842-ba6b-4f84-bff6-054b7292d82e\") " pod="openstack/nova-cell1-91eb-account-create-update-59l2z" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.208046 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nng29\" (UniqueName: \"kubernetes.io/projected/75bff842-ba6b-4f84-bff6-054b7292d82e-kube-api-access-nng29\") pod \"nova-cell1-91eb-account-create-update-59l2z\" (UID: \"75bff842-ba6b-4f84-bff6-054b7292d82e\") " pod="openstack/nova-cell1-91eb-account-create-update-59l2z" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.280391 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3101-account-create-update-g8j4t" Dec 03 17:37:16 crc kubenswrapper[4787]: I1203 17:37:16.301882 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-91eb-account-create-update-59l2z" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.824069 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5bcbf94467-l2njc"] Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.855101 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-75d5f9cb88-48rbj"] Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.885217 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-5c7847c888-455sv"] Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.886644 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.902343 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-internal-svc" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.902566 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-api-public-svc" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.902596 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-5566764dd4-qqp6l"] Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.903950 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.909115 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-internal-svc" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.909323 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-heat-cfnapi-public-svc" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.921081 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5566764dd4-qqp6l"] Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.931487 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-internal-tls-certs\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.931541 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-public-tls-certs\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.931575 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-public-tls-certs\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.931637 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nfzw\" (UniqueName: \"kubernetes.io/projected/21511f7a-7ed3-48ac-aaf9-3580fbc19111-kube-api-access-2nfzw\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.931896 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-combined-ca-bundle\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.933846 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data-custom\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.933940 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.933958 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hr8l9\" (UniqueName: \"kubernetes.io/projected/92c6dbac-064e-484d-aeee-65678e15863a-kube-api-access-hr8l9\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.934004 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-combined-ca-bundle\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.934068 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-internal-tls-certs\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.934119 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.934143 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data-custom\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:17 crc kubenswrapper[4787]: I1203 17:37:17.951826 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5c7847c888-455sv"] Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035462 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nfzw\" (UniqueName: \"kubernetes.io/projected/21511f7a-7ed3-48ac-aaf9-3580fbc19111-kube-api-access-2nfzw\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035508 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-combined-ca-bundle\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035538 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data-custom\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035573 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035589 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hr8l9\" (UniqueName: \"kubernetes.io/projected/92c6dbac-064e-484d-aeee-65678e15863a-kube-api-access-hr8l9\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035616 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-combined-ca-bundle\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035640 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-internal-tls-certs\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035666 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035681 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data-custom\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035724 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-internal-tls-certs\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035754 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-public-tls-certs\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.035781 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-public-tls-certs\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.060464 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-internal-tls-certs\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.091555 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-public-tls-certs\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.102375 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-internal-tls-certs\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.107559 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-combined-ca-bundle\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.108694 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-public-tls-certs\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.109126 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data-custom\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.109825 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.111471 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hr8l9\" (UniqueName: \"kubernetes.io/projected/92c6dbac-064e-484d-aeee-65678e15863a-kube-api-access-hr8l9\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.111569 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data-custom\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.121254 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.122063 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nfzw\" (UniqueName: \"kubernetes.io/projected/21511f7a-7ed3-48ac-aaf9-3580fbc19111-kube-api-access-2nfzw\") pod \"heat-api-5c7847c888-455sv\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.123004 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-combined-ca-bundle\") pod \"heat-cfnapi-5566764dd4-qqp6l\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.245196 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.265447 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:18 crc kubenswrapper[4787]: E1203 17:37:18.778578 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified" Dec 03 17:37:18 crc kubenswrapper[4787]: E1203 17:37:18.778789 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:openstackclient,Image:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,Command:[/bin/sleep],Args:[infinity],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n69h549h68fh676h5d9hb9h78h56h5d5h6bh5b5h6fh56dh67h55chcdhc4h599hcfh647h65bhcdh578h589h5b7hd8h666h68h67bh7dh548h5fbq,ValueFrom:nil,},EnvVar{Name:OS_CLOUD,Value:default,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_CA_CERT,Value:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_HOST,Value:metric-storage-prometheus.openstack.svc,ValueFrom:nil,},EnvVar{Name:PROMETHEUS_PORT,Value:9090,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:openstack-config,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/home/cloud-admin/cloudrc,SubPath:cloudrc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gcf2p,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42401,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42401,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstackclient_openstack(ff54f756-3848-48ec-a235-d2814ff8d7f8): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:37:18 crc kubenswrapper[4787]: E1203 17:37:18.780045 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/openstackclient" podUID="ff54f756-3848-48ec-a235-d2814ff8d7f8" Dec 03 17:37:18 crc kubenswrapper[4787]: W1203 17:37:18.846575 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc14f381_14f6_4719_a0a9_69bfc737fcb5.slice/crio-e51c8d5dcb395a26e64db3d2bbfe9fba858f9d1312e51c473fc997b5ea75f108 WatchSource:0}: Error finding container e51c8d5dcb395a26e64db3d2bbfe9fba858f9d1312e51c473fc997b5ea75f108: Status 404 returned error can't find the container with id e51c8d5dcb395a26e64db3d2bbfe9fba858f9d1312e51c473fc997b5ea75f108 Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.989782 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:37:18 crc kubenswrapper[4787]: I1203 17:37:18.989824 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:37:19 crc kubenswrapper[4787]: I1203 17:37:19.836619 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:37:19 crc kubenswrapper[4787]: I1203 17:37:19.846833 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-l98hp" event={"ID":"1a553bba-383c-4507-bf61-43363d87959b","Type":"ContainerDied","Data":"312c96a6c67f650c683373dcadc929669d5e0248c57f8c644911056358593733"} Dec 03 17:37:19 crc kubenswrapper[4787]: I1203 17:37:19.846881 4787 scope.go:117] "RemoveContainer" containerID="49faf87cf76ca219b67501fd02a55ff8a95c2185cdff111b6d5ae9b9a8a8b9ab" Dec 03 17:37:19 crc kubenswrapper[4787]: I1203 17:37:19.851285 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5bcbf94467-l2njc" event={"ID":"dc14f381-14f6-4719-a0a9-69bfc737fcb5","Type":"ContainerStarted","Data":"e51c8d5dcb395a26e64db3d2bbfe9fba858f9d1312e51c473fc997b5ea75f108"} Dec 03 17:37:19 crc kubenswrapper[4787]: E1203 17:37:19.866722 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"openstackclient\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified\\\"\"" pod="openstack/openstackclient" podUID="ff54f756-3848-48ec-a235-d2814ff8d7f8" Dec 03 17:37:19 crc kubenswrapper[4787]: I1203 17:37:19.901943 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:37:19 crc kubenswrapper[4787]: I1203 17:37:19.953420 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:19 crc kubenswrapper[4787]: I1203 17:37:19.976868 4787 scope.go:117] "RemoveContainer" containerID="151d9d02ef4416fa639752d1217ec7ca9801ed76dad2150a9945371330ae8255" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.000336 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vmhrf\" (UniqueName: \"kubernetes.io/projected/1a553bba-383c-4507-bf61-43363d87959b-kube-api-access-vmhrf\") pod \"1a553bba-383c-4507-bf61-43363d87959b\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.000422 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-ovndb-tls-certs\") pod \"a5a47435-c445-47ff-8806-aa422090813c\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.000451 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-catalog-content\") pod \"1a553bba-383c-4507-bf61-43363d87959b\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.000548 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-config\") pod \"a5a47435-c445-47ff-8806-aa422090813c\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.000610 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-combined-ca-bundle\") pod \"a5a47435-c445-47ff-8806-aa422090813c\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.000661 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-httpd-config\") pod \"a5a47435-c445-47ff-8806-aa422090813c\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.000699 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4j7pb\" (UniqueName: \"kubernetes.io/projected/a5a47435-c445-47ff-8806-aa422090813c-kube-api-access-4j7pb\") pod \"a5a47435-c445-47ff-8806-aa422090813c\" (UID: \"a5a47435-c445-47ff-8806-aa422090813c\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.000736 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-utilities\") pod \"1a553bba-383c-4507-bf61-43363d87959b\" (UID: \"1a553bba-383c-4507-bf61-43363d87959b\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.010567 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-utilities" (OuterVolumeSpecName: "utilities") pod "1a553bba-383c-4507-bf61-43363d87959b" (UID: "1a553bba-383c-4507-bf61-43363d87959b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.022546 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a553bba-383c-4507-bf61-43363d87959b-kube-api-access-vmhrf" (OuterVolumeSpecName: "kube-api-access-vmhrf") pod "1a553bba-383c-4507-bf61-43363d87959b" (UID: "1a553bba-383c-4507-bf61-43363d87959b"). InnerVolumeSpecName "kube-api-access-vmhrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.047347 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5a47435-c445-47ff-8806-aa422090813c-kube-api-access-4j7pb" (OuterVolumeSpecName: "kube-api-access-4j7pb") pod "a5a47435-c445-47ff-8806-aa422090813c" (UID: "a5a47435-c445-47ff-8806-aa422090813c"). InnerVolumeSpecName "kube-api-access-4j7pb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.050867 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "a5a47435-c445-47ff-8806-aa422090813c" (UID: "a5a47435-c445-47ff-8806-aa422090813c"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.092387 4787 scope.go:117] "RemoveContainer" containerID="08fbfdb08e49786a03ebfb4dd9a96224df91ef9566fc1276333133242e4227bc" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.102497 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-combined-ca-bundle\") pod \"fb3be536-b558-4a55-b1ce-f20c67948e10\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.102551 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-scripts\") pod \"fb3be536-b558-4a55-b1ce-f20c67948e10\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.102681 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-log-httpd\") pod \"fb3be536-b558-4a55-b1ce-f20c67948e10\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.102723 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-config-data\") pod \"fb3be536-b558-4a55-b1ce-f20c67948e10\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.102759 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-sg-core-conf-yaml\") pod \"fb3be536-b558-4a55-b1ce-f20c67948e10\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.102894 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mmvqg\" (UniqueName: \"kubernetes.io/projected/fb3be536-b558-4a55-b1ce-f20c67948e10-kube-api-access-mmvqg\") pod \"fb3be536-b558-4a55-b1ce-f20c67948e10\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.102918 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-run-httpd\") pod \"fb3be536-b558-4a55-b1ce-f20c67948e10\" (UID: \"fb3be536-b558-4a55-b1ce-f20c67948e10\") " Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.103504 4787 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.103516 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4j7pb\" (UniqueName: \"kubernetes.io/projected/a5a47435-c445-47ff-8806-aa422090813c-kube-api-access-4j7pb\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.103526 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.103535 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vmhrf\" (UniqueName: \"kubernetes.io/projected/1a553bba-383c-4507-bf61-43363d87959b-kube-api-access-vmhrf\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.105455 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "fb3be536-b558-4a55-b1ce-f20c67948e10" (UID: "fb3be536-b558-4a55-b1ce-f20c67948e10"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.105654 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "fb3be536-b558-4a55-b1ce-f20c67948e10" (UID: "fb3be536-b558-4a55-b1ce-f20c67948e10"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.120147 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb3be536-b558-4a55-b1ce-f20c67948e10-kube-api-access-mmvqg" (OuterVolumeSpecName: "kube-api-access-mmvqg") pod "fb3be536-b558-4a55-b1ce-f20c67948e10" (UID: "fb3be536-b558-4a55-b1ce-f20c67948e10"). InnerVolumeSpecName "kube-api-access-mmvqg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.120719 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-scripts" (OuterVolumeSpecName: "scripts") pod "fb3be536-b558-4a55-b1ce-f20c67948e10" (UID: "fb3be536-b558-4a55-b1ce-f20c67948e10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.146479 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1a553bba-383c-4507-bf61-43363d87959b" (UID: "1a553bba-383c-4507-bf61-43363d87959b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.190340 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5a47435-c445-47ff-8806-aa422090813c" (UID: "a5a47435-c445-47ff-8806-aa422090813c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.205454 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.205503 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.205518 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.205532 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1a553bba-383c-4507-bf61-43363d87959b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.205546 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mmvqg\" (UniqueName: \"kubernetes.io/projected/fb3be536-b558-4a55-b1ce-f20c67948e10-kube-api-access-mmvqg\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.205560 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fb3be536-b558-4a55-b1ce-f20c67948e10-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.209194 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-config" (OuterVolumeSpecName: "config") pod "a5a47435-c445-47ff-8806-aa422090813c" (UID: "a5a47435-c445-47ff-8806-aa422090813c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.249897 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "a5a47435-c445-47ff-8806-aa422090813c" (UID: "a5a47435-c445-47ff-8806-aa422090813c"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.261171 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "fb3be536-b558-4a55-b1ce-f20c67948e10" (UID: "fb3be536-b558-4a55-b1ce-f20c67948e10"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.316404 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.316429 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.316439 4787 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a5a47435-c445-47ff-8806-aa422090813c-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.383349 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-config-data" (OuterVolumeSpecName: "config-data") pod "fb3be536-b558-4a55-b1ce-f20c67948e10" (UID: "fb3be536-b558-4a55-b1ce-f20c67948e10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.383401 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fb3be536-b558-4a55-b1ce-f20c67948e10" (UID: "fb3be536-b558-4a55-b1ce-f20c67948e10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.418803 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.418837 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fb3be536-b558-4a55-b1ce-f20c67948e10-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.462471 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-a7a8-account-create-update-72nlv"] Dec 03 17:37:20 crc kubenswrapper[4787]: W1203 17:37:20.529441 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod39227401_0c97_4fd7_8510_615e22ca73d9.slice/crio-e46ae821f56778655d3267f23ee284cce4c9a40bff63e0afa034ecdb68b05ef6 WatchSource:0}: Error finding container e46ae821f56778655d3267f23ee284cce4c9a40bff63e0afa034ecdb68b05ef6: Status 404 returned error can't find the container with id e46ae821f56778655d3267f23ee284cce4c9a40bff63e0afa034ecdb68b05ef6 Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.920173 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-69d9b47bb8-6fqn7" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.921823 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-69d9b47bb8-6fqn7" event={"ID":"a5a47435-c445-47ff-8806-aa422090813c","Type":"ContainerDied","Data":"5aa581fdb2538a587a267514a9d64b67c51264c7b7f1bb48222d64709bf3175d"} Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.922331 4787 scope.go:117] "RemoveContainer" containerID="6a933cc3673f14caef8f0be371decbf1bd9d8f5be6bcd9b1e80838a264078a2a" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.936971 4787 generic.go:334] "Generic (PLEG): container finished" podID="b67a4816-85d5-404a-a22d-6cfbc455d286" containerID="b76ae34ba03010e44743be5f54b8fd5010504ad6a5669fee127928bb6bd8634c" exitCode=0 Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.937097 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" event={"ID":"b67a4816-85d5-404a-a22d-6cfbc455d286","Type":"ContainerDied","Data":"b76ae34ba03010e44743be5f54b8fd5010504ad6a5669fee127928bb6bd8634c"} Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.951064 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"fb3be536-b558-4a55-b1ce-f20c67948e10","Type":"ContainerDied","Data":"dd15234213bdddd73e5ed84c7af849bbb8f882c17d30e9d36a7ffd4db54cbc1f"} Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.951382 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.975187 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" event={"ID":"39227401-0c97-4fd7-8510-615e22ca73d9","Type":"ContainerStarted","Data":"f88de925b2be6062cba8c51bb2e7a43a3762824e7bcff3dc75c995498933e662"} Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.975236 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" event={"ID":"39227401-0c97-4fd7-8510-615e22ca73d9","Type":"ContainerStarted","Data":"e46ae821f56778655d3267f23ee284cce4c9a40bff63e0afa034ecdb68b05ef6"} Dec 03 17:37:20 crc kubenswrapper[4787]: I1203 17:37:20.982079 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-l98hp" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.006065 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" podStartSLOduration=6.006045366 podStartE2EDuration="6.006045366s" podCreationTimestamp="2025-12-03 17:37:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:37:20.991258195 +0000 UTC m=+1457.808729164" watchObservedRunningTime="2025-12-03 17:37:21.006045366 +0000 UTC m=+1457.823516325" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.062944 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.085430 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.097091 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-l98hp"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.106811 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-l98hp"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.124982 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:21 crc kubenswrapper[4787]: E1203 17:37:21.125559 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="sg-core" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125577 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="sg-core" Dec 03 17:37:21 crc kubenswrapper[4787]: E1203 17:37:21.125615 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a47435-c445-47ff-8806-aa422090813c" containerName="neutron-httpd" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125633 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a47435-c445-47ff-8806-aa422090813c" containerName="neutron-httpd" Dec 03 17:37:21 crc kubenswrapper[4787]: E1203 17:37:21.125652 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5a47435-c445-47ff-8806-aa422090813c" containerName="neutron-api" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125659 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5a47435-c445-47ff-8806-aa422090813c" containerName="neutron-api" Dec 03 17:37:21 crc kubenswrapper[4787]: E1203 17:37:21.125676 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="proxy-httpd" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125683 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="proxy-httpd" Dec 03 17:37:21 crc kubenswrapper[4787]: E1203 17:37:21.125695 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a553bba-383c-4507-bf61-43363d87959b" containerName="registry-server" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125701 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a553bba-383c-4507-bf61-43363d87959b" containerName="registry-server" Dec 03 17:37:21 crc kubenswrapper[4787]: E1203 17:37:21.125712 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a553bba-383c-4507-bf61-43363d87959b" containerName="extract-content" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125718 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a553bba-383c-4507-bf61-43363d87959b" containerName="extract-content" Dec 03 17:37:21 crc kubenswrapper[4787]: E1203 17:37:21.125726 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="ceilometer-central-agent" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125732 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="ceilometer-central-agent" Dec 03 17:37:21 crc kubenswrapper[4787]: E1203 17:37:21.125740 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="ceilometer-notification-agent" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125745 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="ceilometer-notification-agent" Dec 03 17:37:21 crc kubenswrapper[4787]: E1203 17:37:21.125754 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a553bba-383c-4507-bf61-43363d87959b" containerName="extract-utilities" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125760 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a553bba-383c-4507-bf61-43363d87959b" containerName="extract-utilities" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125955 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="ceilometer-central-agent" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125973 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="sg-core" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.125983 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5a47435-c445-47ff-8806-aa422090813c" containerName="neutron-httpd" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.126000 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a553bba-383c-4507-bf61-43363d87959b" containerName="registry-server" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.126029 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="ceilometer-notification-agent" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.126037 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5a47435-c445-47ff-8806-aa422090813c" containerName="neutron-api" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.126048 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" containerName="proxy-httpd" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.130454 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.139447 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.139703 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.189824 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.242370 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-69d9b47bb8-6fqn7"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.264098 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-scripts\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.264164 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-log-httpd\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.264217 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.264467 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g2c5d\" (UniqueName: \"kubernetes.io/projected/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-kube-api-access-g2c5d\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.264633 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.264792 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-config-data\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.264860 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-run-httpd\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.269275 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-69d9b47bb8-6fqn7"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.285694 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.306392 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-91eb-account-create-update-59l2z"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.321255 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-hj85k"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.333774 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-7994c4bbc4-2kvbc"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.355134 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-6dhs9"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.367294 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-65b7898d75-fjwnb"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.369789 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g2c5d\" (UniqueName: \"kubernetes.io/projected/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-kube-api-access-g2c5d\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.369895 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.369982 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-config-data\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.370041 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-run-httpd\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.370123 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-scripts\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.370853 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-run-httpd\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.371262 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-log-httpd\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.371349 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.371852 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-log-httpd\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.381816 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.385744 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-scripts\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.386042 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.386928 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-config-data\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.389180 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g2c5d\" (UniqueName: \"kubernetes.io/projected/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-kube-api-access-g2c5d\") pod \"ceilometer-0\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.498697 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.613107 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-64cb8b689b-d4pmz"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.626421 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-lcm5g"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.636696 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-5c7847c888-455sv"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.647178 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-5566764dd4-qqp6l"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.658632 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-3101-account-create-update-g8j4t"] Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.673392 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-8b787c4b7-wwr58"] Dec 03 17:37:21 crc kubenswrapper[4787]: W1203 17:37:21.754250 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod92c6dbac_064e_484d_aeee_65678e15863a.slice/crio-da20ced467ef59ed0e26b364b2423ad3276a47445a969a983cc58be02e811b46 WatchSource:0}: Error finding container da20ced467ef59ed0e26b364b2423ad3276a47445a969a983cc58be02e811b46: Status 404 returned error can't find the container with id da20ced467ef59ed0e26b364b2423ad3276a47445a969a983cc58be02e811b46 Dec 03 17:37:21 crc kubenswrapper[4787]: W1203 17:37:21.759042 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21511f7a_7ed3_48ac_aaf9_3580fbc19111.slice/crio-aaa229a64996c290c603b07fed035883e4972f344c980a87e002f4f808fdd7e5 WatchSource:0}: Error finding container aaa229a64996c290c603b07fed035883e4972f344c980a87e002f4f808fdd7e5: Status 404 returned error can't find the container with id aaa229a64996c290c603b07fed035883e4972f344c980a87e002f4f808fdd7e5 Dec 03 17:37:21 crc kubenswrapper[4787]: W1203 17:37:21.772656 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod244e40a3_b432_4971_bdf5_f070ab9891b0.slice/crio-5b08eb4f7625d13fe465489f5edd5af57e7f41abbcd4daa9f406bca990cf4a02 WatchSource:0}: Error finding container 5b08eb4f7625d13fe465489f5edd5af57e7f41abbcd4daa9f406bca990cf4a02: Status 404 returned error can't find the container with id 5b08eb4f7625d13fe465489f5edd5af57e7f41abbcd4daa9f406bca990cf4a02 Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.800544 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a553bba-383c-4507-bf61-43363d87959b" path="/var/lib/kubelet/pods/1a553bba-383c-4507-bf61-43363d87959b/volumes" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.801232 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5a47435-c445-47ff-8806-aa422090813c" path="/var/lib/kubelet/pods/a5a47435-c445-47ff-8806-aa422090813c/volumes" Dec 03 17:37:21 crc kubenswrapper[4787]: I1203 17:37:21.801830 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb3be536-b558-4a55-b1ce-f20c67948e10" path="/var/lib/kubelet/pods/fb3be536-b558-4a55-b1ce-f20c67948e10/volumes" Dec 03 17:37:22 crc kubenswrapper[4787]: I1203 17:37:22.005152 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-lcm5g" event={"ID":"244e40a3-b432-4971-bdf5-f070ab9891b0","Type":"ContainerStarted","Data":"5b08eb4f7625d13fe465489f5edd5af57e7f41abbcd4daa9f406bca990cf4a02"} Dec 03 17:37:22 crc kubenswrapper[4787]: I1203 17:37:22.010952 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5c7847c888-455sv" event={"ID":"21511f7a-7ed3-48ac-aaf9-3580fbc19111","Type":"ContainerStarted","Data":"aaa229a64996c290c603b07fed035883e4972f344c980a87e002f4f808fdd7e5"} Dec 03 17:37:22 crc kubenswrapper[4787]: I1203 17:37:22.013689 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7994c4bbc4-2kvbc" event={"ID":"9296f967-7bf3-4e42-a867-cce34092dc11","Type":"ContainerStarted","Data":"84bbcb5aaef296b8c23267f71f9f80c12f357655836ee1e36d97193d0e414cc3"} Dec 03 17:37:22 crc kubenswrapper[4787]: I1203 17:37:22.018601 4787 generic.go:334] "Generic (PLEG): container finished" podID="39227401-0c97-4fd7-8510-615e22ca73d9" containerID="f88de925b2be6062cba8c51bb2e7a43a3762824e7bcff3dc75c995498933e662" exitCode=0 Dec 03 17:37:22 crc kubenswrapper[4787]: I1203 17:37:22.018691 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" event={"ID":"39227401-0c97-4fd7-8510-615e22ca73d9","Type":"ContainerDied","Data":"f88de925b2be6062cba8c51bb2e7a43a3762824e7bcff3dc75c995498933e662"} Dec 03 17:37:22 crc kubenswrapper[4787]: I1203 17:37:22.020175 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3101-account-create-update-g8j4t" event={"ID":"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a","Type":"ContainerStarted","Data":"32a32fb37176489b55bbf6d661a992c81cf0069e793d28ac2a69ba61336050ea"} Dec 03 17:37:22 crc kubenswrapper[4787]: I1203 17:37:22.023486 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" event={"ID":"92c6dbac-064e-484d-aeee-65678e15863a","Type":"ContainerStarted","Data":"da20ced467ef59ed0e26b364b2423ad3276a47445a969a983cc58be02e811b46"} Dec 03 17:37:22 crc kubenswrapper[4787]: W1203 17:37:22.271399 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1a23593c_7762_4639_8b2f_2cdd49b0c728.slice/crio-e9d7997a08de7b48222229340121e1186628293d7f3db99b553e932840f8606f WatchSource:0}: Error finding container e9d7997a08de7b48222229340121e1186628293d7f3db99b553e932840f8606f: Status 404 returned error can't find the container with id e9d7997a08de7b48222229340121e1186628293d7f3db99b553e932840f8606f Dec 03 17:37:22 crc kubenswrapper[4787]: W1203 17:37:22.276568 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cb81af5_f082_4509_b5a5_b0a49301b75e.slice/crio-b9d3b117a2a90f24336c569c59fb487a244db51e61afe1f3109d2d0545ce1a94 WatchSource:0}: Error finding container b9d3b117a2a90f24336c569c59fb487a244db51e61afe1f3109d2d0545ce1a94: Status 404 returned error can't find the container with id b9d3b117a2a90f24336c569c59fb487a244db51e61afe1f3109d2d0545ce1a94 Dec 03 17:37:22 crc kubenswrapper[4787]: W1203 17:37:22.289743 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f12c699_048b_4eee_bc7c_e3ad916656ed.slice/crio-6b1ad6375492c45a59dc60c5904a65d3dd7e8f5b584633cd82fae68218449a4e WatchSource:0}: Error finding container 6b1ad6375492c45a59dc60c5904a65d3dd7e8f5b584633cd82fae68218449a4e: Status 404 returned error can't find the container with id 6b1ad6375492c45a59dc60c5904a65d3dd7e8f5b584633cd82fae68218449a4e Dec 03 17:37:22 crc kubenswrapper[4787]: I1203 17:37:22.478553 4787 scope.go:117] "RemoveContainer" containerID="fb23933ee3023e829dbcbf4c0a0602af977cacafc66c805de1fff16c6ca91111" Dec 03 17:37:22 crc kubenswrapper[4787]: I1203 17:37:22.933100 4787 scope.go:117] "RemoveContainer" containerID="ebbca1e662829b7c30611980184f208b1c1dac07f7e369d55f6016eab924f980" Dec 03 17:37:23 crc kubenswrapper[4787]: I1203 17:37:23.041393 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6dhs9" event={"ID":"1a23593c-7762-4639-8b2f-2cdd49b0c728","Type":"ContainerStarted","Data":"e9d7997a08de7b48222229340121e1186628293d7f3db99b553e932840f8606f"} Dec 03 17:37:23 crc kubenswrapper[4787]: I1203 17:37:23.043976 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" event={"ID":"8f12c699-048b-4eee-bc7c-e3ad916656ed","Type":"ContainerStarted","Data":"6b1ad6375492c45a59dc60c5904a65d3dd7e8f5b584633cd82fae68218449a4e"} Dec 03 17:37:23 crc kubenswrapper[4787]: I1203 17:37:23.049609 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hj85k" event={"ID":"6371d53b-b035-4a6e-82b5-c5774ffc8de1","Type":"ContainerStarted","Data":"d0f135984c8ebe69810235a15d965cad8f9fa74d63fafb683d170f6bc9322bd4"} Dec 03 17:37:23 crc kubenswrapper[4787]: I1203 17:37:23.058611 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9cb81af5-f082-4509-b5a5-b0a49301b75e","Type":"ContainerStarted","Data":"b9d3b117a2a90f24336c569c59fb487a244db51e61afe1f3109d2d0545ce1a94"} Dec 03 17:37:23 crc kubenswrapper[4787]: I1203 17:37:23.066316 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-8b787c4b7-wwr58" event={"ID":"fef2cdf6-1a2c-4904-9010-c446ad96c3a7","Type":"ContainerStarted","Data":"0817b80698ab5b085607d874587c272e877277250405a26ed23b7f1ce993348a"} Dec 03 17:37:23 crc kubenswrapper[4787]: I1203 17:37:23.094044 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:23 crc kubenswrapper[4787]: I1203 17:37:23.096498 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65b7898d75-fjwnb" event={"ID":"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee","Type":"ContainerStarted","Data":"264e89170c8ba95d368ee0046c5c17e8c7b873589804a1304efda0384f9975e8"} Dec 03 17:37:23 crc kubenswrapper[4787]: I1203 17:37:23.097446 4787 scope.go:117] "RemoveContainer" containerID="c152187ae65339c6eb3bd61300787d5a2150b6dbc472213946c01ec22db54d15" Dec 03 17:37:23 crc kubenswrapper[4787]: I1203 17:37:23.098758 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-91eb-account-create-update-59l2z" event={"ID":"75bff842-ba6b-4f84-bff6-054b7292d82e","Type":"ContainerStarted","Data":"60aec9976e5ab8829cde63b74025abc8da5dcf4a7e9952444eaf0709c0323e9a"} Dec 03 17:37:23 crc kubenswrapper[4787]: I1203 17:37:23.132087 4787 scope.go:117] "RemoveContainer" containerID="e863c048d1f5611690721248ddef6735980514cd16d0d6ae882ee94c2a0998de" Dec 03 17:37:23 crc kubenswrapper[4787]: I1203 17:37:23.195509 4787 scope.go:117] "RemoveContainer" containerID="536728db935e62e50d81a2c5b98d7a76d3b39fe90f7536a4819eef30f73bfd0f" Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.028749 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.115834 4787 generic.go:334] "Generic (PLEG): container finished" podID="75bff842-ba6b-4f84-bff6-054b7292d82e" containerID="094e601df92d88190a1f624d9ac4d8030d1e73bd06959edc6e27b20fe473f3e2" exitCode=0 Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.115919 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-91eb-account-create-update-59l2z" event={"ID":"75bff842-ba6b-4f84-bff6-054b7292d82e","Type":"ContainerDied","Data":"094e601df92d88190a1f624d9ac4d8030d1e73bd06959edc6e27b20fe473f3e2"} Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.120483 4787 generic.go:334] "Generic (PLEG): container finished" podID="2ea7d1a3-de18-4283-bd8e-2be80f82eb9a" containerID="32cf70c8718d0cf030ccd84af5935a7264a1072b0d92369eb54dfd538ddedf04" exitCode=0 Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.120629 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3101-account-create-update-g8j4t" event={"ID":"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a","Type":"ContainerDied","Data":"32cf70c8718d0cf030ccd84af5935a7264a1072b0d92369eb54dfd538ddedf04"} Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.123602 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" event={"ID":"b67a4816-85d5-404a-a22d-6cfbc455d286","Type":"ContainerStarted","Data":"b2c6b8eb60b0320b23d2a30b6c8301c807adc10713a22a67b0f04a7e88b4bbcf"} Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.124263 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.125614 4787 generic.go:334] "Generic (PLEG): container finished" podID="244e40a3-b432-4971-bdf5-f070ab9891b0" containerID="14e0483968ab6ca747903bdaf85df573297632c5106490083f10f40a4fbd5d88" exitCode=0 Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.125681 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-lcm5g" event={"ID":"244e40a3-b432-4971-bdf5-f070ab9891b0","Type":"ContainerDied","Data":"14e0483968ab6ca747903bdaf85df573297632c5106490083f10f40a4fbd5d88"} Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.128078 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" event={"ID":"39227401-0c97-4fd7-8510-615e22ca73d9","Type":"ContainerDied","Data":"e46ae821f56778655d3267f23ee284cce4c9a40bff63e0afa034ecdb68b05ef6"} Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.128110 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e46ae821f56778655d3267f23ee284cce4c9a40bff63e0afa034ecdb68b05ef6" Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.128123 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-a7a8-account-create-update-72nlv" Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.132068 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"956e5a96-03a5-40ca-a9ce-fbdc8631edb3","Type":"ContainerStarted","Data":"7700d42651bc8bd214ea8f1fb2ff339b3a67dca1566dc6dc7ab5a224ee680d2c"} Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.133417 4787 generic.go:334] "Generic (PLEG): container finished" podID="6371d53b-b035-4a6e-82b5-c5774ffc8de1" containerID="5f8bd0c2316a0b41c49138685d20b95d4b5a8b7d5ba8721f05f2b058295c00e6" exitCode=0 Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.133479 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hj85k" event={"ID":"6371d53b-b035-4a6e-82b5-c5774ffc8de1","Type":"ContainerDied","Data":"5f8bd0c2316a0b41c49138685d20b95d4b5a8b7d5ba8721f05f2b058295c00e6"} Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.135756 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-8b787c4b7-wwr58" event={"ID":"fef2cdf6-1a2c-4904-9010-c446ad96c3a7","Type":"ContainerStarted","Data":"437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb"} Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.136561 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.139790 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39227401-0c97-4fd7-8510-615e22ca73d9-operator-scripts\") pod \"39227401-0c97-4fd7-8510-615e22ca73d9\" (UID: \"39227401-0c97-4fd7-8510-615e22ca73d9\") " Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.139855 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ztbw9\" (UniqueName: \"kubernetes.io/projected/39227401-0c97-4fd7-8510-615e22ca73d9-kube-api-access-ztbw9\") pod \"39227401-0c97-4fd7-8510-615e22ca73d9\" (UID: \"39227401-0c97-4fd7-8510-615e22ca73d9\") " Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.140535 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/39227401-0c97-4fd7-8510-615e22ca73d9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "39227401-0c97-4fd7-8510-615e22ca73d9" (UID: "39227401-0c97-4fd7-8510-615e22ca73d9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.147822 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/39227401-0c97-4fd7-8510-615e22ca73d9-kube-api-access-ztbw9" (OuterVolumeSpecName: "kube-api-access-ztbw9") pod "39227401-0c97-4fd7-8510-615e22ca73d9" (UID: "39227401-0c97-4fd7-8510-615e22ca73d9"). InnerVolumeSpecName "kube-api-access-ztbw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.194987 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" podStartSLOduration=18.194968372 podStartE2EDuration="18.194968372s" podCreationTimestamp="2025-12-03 17:37:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:37:24.187983243 +0000 UTC m=+1461.005454212" watchObservedRunningTime="2025-12-03 17:37:24.194968372 +0000 UTC m=+1461.012439321" Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.225960 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-8b787c4b7-wwr58" podStartSLOduration=9.2259389 podStartE2EDuration="9.2259389s" podCreationTimestamp="2025-12-03 17:37:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:37:24.225358784 +0000 UTC m=+1461.042829743" watchObservedRunningTime="2025-12-03 17:37:24.2259389 +0000 UTC m=+1461.043409859" Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.243063 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/39227401-0c97-4fd7-8510-615e22ca73d9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:24 crc kubenswrapper[4787]: I1203 17:37:24.243094 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ztbw9\" (UniqueName: \"kubernetes.io/projected/39227401-0c97-4fd7-8510-615e22ca73d9-kube-api-access-ztbw9\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.171418 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"956e5a96-03a5-40ca-a9ce-fbdc8631edb3","Type":"ContainerStarted","Data":"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a"} Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.174743 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-75d5f9cb88-48rbj" event={"ID":"455e1ae0-75b9-4c80-9580-7b2ba4c72483","Type":"ContainerStarted","Data":"030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5"} Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.174925 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-75d5f9cb88-48rbj" podUID="455e1ae0-75b9-4c80-9580-7b2ba4c72483" containerName="heat-api" containerID="cri-o://030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5" gracePeriod=60 Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.175258 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.204824 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-75d5f9cb88-48rbj" podStartSLOduration=4.370369092 podStartE2EDuration="19.204800839s" podCreationTimestamp="2025-12-03 17:37:06 +0000 UTC" firstStartedPulling="2025-12-03 17:37:07.735339129 +0000 UTC m=+1444.552810078" lastFinishedPulling="2025-12-03 17:37:22.569770866 +0000 UTC m=+1459.387241825" observedRunningTime="2025-12-03 17:37:25.199897616 +0000 UTC m=+1462.017368575" watchObservedRunningTime="2025-12-03 17:37:25.204800839 +0000 UTC m=+1462.022271798" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.210286 4787 generic.go:334] "Generic (PLEG): container finished" podID="1a23593c-7762-4639-8b2f-2cdd49b0c728" containerID="4db0b4029701f28ef38c1a9528668be4f59569f70282bda4cc0240c8f55fcb83" exitCode=0 Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.210910 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6dhs9" event={"ID":"1a23593c-7762-4639-8b2f-2cdd49b0c728","Type":"ContainerDied","Data":"4db0b4029701f28ef38c1a9528668be4f59569f70282bda4cc0240c8f55fcb83"} Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.224306 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9cb81af5-f082-4509-b5a5-b0a49301b75e","Type":"ContainerStarted","Data":"1b86fdc2d699cc9f4f6450341b650b2cb45689f803af9e49e930fb27f3b05eb4"} Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.230774 4787 generic.go:334] "Generic (PLEG): container finished" podID="9296f967-7bf3-4e42-a867-cce34092dc11" containerID="23f7168ce7eebdba833bc2c4cb8e6e1b44620b162f312222dbde84b090254536" exitCode=1 Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.231065 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7994c4bbc4-2kvbc" event={"ID":"9296f967-7bf3-4e42-a867-cce34092dc11","Type":"ContainerDied","Data":"23f7168ce7eebdba833bc2c4cb8e6e1b44620b162f312222dbde84b090254536"} Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.249558 4787 scope.go:117] "RemoveContainer" containerID="23f7168ce7eebdba833bc2c4cb8e6e1b44620b162f312222dbde84b090254536" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.277406 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5bcbf94467-l2njc" event={"ID":"dc14f381-14f6-4719-a0a9-69bfc737fcb5","Type":"ContainerStarted","Data":"4786ee5b7941d5e9659f45efe86cd1ca5e32ea0024e2d05b1cd3908ec5f53202"} Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.277591 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-5bcbf94467-l2njc" podUID="dc14f381-14f6-4719-a0a9-69bfc737fcb5" containerName="heat-cfnapi" containerID="cri-o://4786ee5b7941d5e9659f45efe86cd1ca5e32ea0024e2d05b1cd3908ec5f53202" gracePeriod=60 Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.278154 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.292881 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" event={"ID":"92c6dbac-064e-484d-aeee-65678e15863a","Type":"ContainerStarted","Data":"4add3899182083f4a01561f2e932c2e4d8bf2672805901f959479a1689ab80a5"} Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.298234 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.340193 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5c7847c888-455sv" event={"ID":"21511f7a-7ed3-48ac-aaf9-3580fbc19111","Type":"ContainerStarted","Data":"f32e2dabb9c3e581fef8022b3a27a1836dce1b3ed6ade71b41d28a8290c08e06"} Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.347728 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.397276 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5bcbf94467-l2njc" podStartSLOduration=15.561929524 podStartE2EDuration="19.397257599s" podCreationTimestamp="2025-12-03 17:37:06 +0000 UTC" firstStartedPulling="2025-12-03 17:37:18.862804366 +0000 UTC m=+1455.680275325" lastFinishedPulling="2025-12-03 17:37:22.698132441 +0000 UTC m=+1459.515603400" observedRunningTime="2025-12-03 17:37:25.304069996 +0000 UTC m=+1462.121541325" watchObservedRunningTime="2025-12-03 17:37:25.397257599 +0000 UTC m=+1462.214728558" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.397763 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65b7898d75-fjwnb" event={"ID":"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee","Type":"ContainerStarted","Data":"6c3310810e0f7b8b5b08498142a9bf3d8aee104dbb3833309e6a62082fec1787"} Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.407351 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" podStartSLOduration=7.06783157 podStartE2EDuration="8.407328311s" podCreationTimestamp="2025-12-03 17:37:17 +0000 UTC" firstStartedPulling="2025-12-03 17:37:21.759167023 +0000 UTC m=+1458.576637982" lastFinishedPulling="2025-12-03 17:37:23.098663764 +0000 UTC m=+1459.916134723" observedRunningTime="2025-12-03 17:37:25.343610736 +0000 UTC m=+1462.161081725" watchObservedRunningTime="2025-12-03 17:37:25.407328311 +0000 UTC m=+1462.224799270" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.407458 4787 generic.go:334] "Generic (PLEG): container finished" podID="8f12c699-048b-4eee-bc7c-e3ad916656ed" containerID="8c5c4776b7eed2d9137585ab9027c862bab6426b3adf863c7ff79b7c0b85ba8e" exitCode=1 Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.408358 4787 scope.go:117] "RemoveContainer" containerID="8c5c4776b7eed2d9137585ab9027c862bab6426b3adf863c7ff79b7c0b85ba8e" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.411796 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" event={"ID":"8f12c699-048b-4eee-bc7c-e3ad916656ed","Type":"ContainerDied","Data":"8c5c4776b7eed2d9137585ab9027c862bab6426b3adf863c7ff79b7c0b85ba8e"} Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.440309 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-5c7847c888-455sv" podStartSLOduration=7.15086247 podStartE2EDuration="8.440289774s" podCreationTimestamp="2025-12-03 17:37:17 +0000 UTC" firstStartedPulling="2025-12-03 17:37:21.763567213 +0000 UTC m=+1458.581038172" lastFinishedPulling="2025-12-03 17:37:23.052994507 +0000 UTC m=+1459.870465476" observedRunningTime="2025-12-03 17:37:25.379550229 +0000 UTC m=+1462.197021188" watchObservedRunningTime="2025-12-03 17:37:25.440289774 +0000 UTC m=+1462.257760733" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.463869 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.463911 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.478533 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:25 crc kubenswrapper[4787]: I1203 17:37:25.478569 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.136503 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hj85k" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.215736 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97mgf\" (UniqueName: \"kubernetes.io/projected/6371d53b-b035-4a6e-82b5-c5774ffc8de1-kube-api-access-97mgf\") pod \"6371d53b-b035-4a6e-82b5-c5774ffc8de1\" (UID: \"6371d53b-b035-4a6e-82b5-c5774ffc8de1\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.216197 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6371d53b-b035-4a6e-82b5-c5774ffc8de1-operator-scripts\") pod \"6371d53b-b035-4a6e-82b5-c5774ffc8de1\" (UID: \"6371d53b-b035-4a6e-82b5-c5774ffc8de1\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.217657 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6371d53b-b035-4a6e-82b5-c5774ffc8de1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6371d53b-b035-4a6e-82b5-c5774ffc8de1" (UID: "6371d53b-b035-4a6e-82b5-c5774ffc8de1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.229238 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6371d53b-b035-4a6e-82b5-c5774ffc8de1-kube-api-access-97mgf" (OuterVolumeSpecName: "kube-api-access-97mgf") pod "6371d53b-b035-4a6e-82b5-c5774ffc8de1" (UID: "6371d53b-b035-4a6e-82b5-c5774ffc8de1"). InnerVolumeSpecName "kube-api-access-97mgf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.319603 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6371d53b-b035-4a6e-82b5-c5774ffc8de1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.319640 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97mgf\" (UniqueName: \"kubernetes.io/projected/6371d53b-b035-4a6e-82b5-c5774ffc8de1-kube-api-access-97mgf\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.406342 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3101-account-create-update-g8j4t" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.410582 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.468738 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-lcm5g" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.480277 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-91eb-account-create-update-59l2z" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.481772 4787 generic.go:334] "Generic (PLEG): container finished" podID="8f12c699-048b-4eee-bc7c-e3ad916656ed" containerID="7e594040c2aee27dd4dac76f8492ed0427c1456eebe85855a32e5f68443f301a" exitCode=1 Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.481899 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" event={"ID":"8f12c699-048b-4eee-bc7c-e3ad916656ed","Type":"ContainerDied","Data":"7e594040c2aee27dd4dac76f8492ed0427c1456eebe85855a32e5f68443f301a"} Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.481936 4787 scope.go:117] "RemoveContainer" containerID="8c5c4776b7eed2d9137585ab9027c862bab6426b3adf863c7ff79b7c0b85ba8e" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.483149 4787 scope.go:117] "RemoveContainer" containerID="7e594040c2aee27dd4dac76f8492ed0427c1456eebe85855a32e5f68443f301a" Dec 03 17:37:26 crc kubenswrapper[4787]: E1203 17:37:26.483542 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-64cb8b689b-d4pmz_openstack(8f12c699-048b-4eee-bc7c-e3ad916656ed)\"" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" podUID="8f12c699-048b-4eee-bc7c-e3ad916656ed" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.505001 4787 generic.go:334] "Generic (PLEG): container finished" podID="d774a24e-5451-4976-a0d5-f915ebe863bf" containerID="e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183" exitCode=137 Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.505087 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d774a24e-5451-4976-a0d5-f915ebe863bf","Type":"ContainerDied","Data":"e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183"} Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.505115 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"d774a24e-5451-4976-a0d5-f915ebe863bf","Type":"ContainerDied","Data":"3c3d748c3f5516fbef788f28e5f8bf6b24546bb4b0f791edb2b95c14ebb3d30c"} Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.505176 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.515373 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"9cb81af5-f082-4509-b5a5-b0a49301b75e","Type":"ContainerStarted","Data":"b86b5c5073c908c006e57b0721d32588b2bdae06944da387ef57257f9b6cd370"} Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.532679 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data\") pod \"d774a24e-5451-4976-a0d5-f915ebe863bf\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.535853 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"956e5a96-03a5-40ca-a9ce-fbdc8631edb3","Type":"ContainerStarted","Data":"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568"} Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.538590 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-combined-ca-bundle\") pod \"d774a24e-5451-4976-a0d5-f915ebe863bf\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.538646 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d774a24e-5451-4976-a0d5-f915ebe863bf-etc-machine-id\") pod \"d774a24e-5451-4976-a0d5-f915ebe863bf\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.538727 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data-custom\") pod \"d774a24e-5451-4976-a0d5-f915ebe863bf\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.538748 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h66vg\" (UniqueName: \"kubernetes.io/projected/d774a24e-5451-4976-a0d5-f915ebe863bf-kube-api-access-h66vg\") pod \"d774a24e-5451-4976-a0d5-f915ebe863bf\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.538799 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbvk7\" (UniqueName: \"kubernetes.io/projected/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-kube-api-access-dbvk7\") pod \"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a\" (UID: \"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.538854 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/244e40a3-b432-4971-bdf5-f070ab9891b0-operator-scripts\") pod \"244e40a3-b432-4971-bdf5-f070ab9891b0\" (UID: \"244e40a3-b432-4971-bdf5-f070ab9891b0\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.538877 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d774a24e-5451-4976-a0d5-f915ebe863bf-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "d774a24e-5451-4976-a0d5-f915ebe863bf" (UID: "d774a24e-5451-4976-a0d5-f915ebe863bf"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.538925 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5b9s9\" (UniqueName: \"kubernetes.io/projected/244e40a3-b432-4971-bdf5-f070ab9891b0-kube-api-access-5b9s9\") pod \"244e40a3-b432-4971-bdf5-f070ab9891b0\" (UID: \"244e40a3-b432-4971-bdf5-f070ab9891b0\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.538956 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-operator-scripts\") pod \"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a\" (UID: \"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.538979 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d774a24e-5451-4976-a0d5-f915ebe863bf-logs\") pod \"d774a24e-5451-4976-a0d5-f915ebe863bf\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.539006 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-scripts\") pod \"d774a24e-5451-4976-a0d5-f915ebe863bf\" (UID: \"d774a24e-5451-4976-a0d5-f915ebe863bf\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.539742 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/244e40a3-b432-4971-bdf5-f070ab9891b0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "244e40a3-b432-4971-bdf5-f070ab9891b0" (UID: "244e40a3-b432-4971-bdf5-f070ab9891b0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.540510 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/244e40a3-b432-4971-bdf5-f070ab9891b0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.549843 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/244e40a3-b432-4971-bdf5-f070ab9891b0-kube-api-access-5b9s9" (OuterVolumeSpecName: "kube-api-access-5b9s9") pod "244e40a3-b432-4971-bdf5-f070ab9891b0" (UID: "244e40a3-b432-4971-bdf5-f070ab9891b0"). InnerVolumeSpecName "kube-api-access-5b9s9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.550317 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2ea7d1a3-de18-4283-bd8e-2be80f82eb9a" (UID: "2ea7d1a3-de18-4283-bd8e-2be80f82eb9a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.550708 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d774a24e-5451-4976-a0d5-f915ebe863bf-logs" (OuterVolumeSpecName: "logs") pod "d774a24e-5451-4976-a0d5-f915ebe863bf" (UID: "d774a24e-5451-4976-a0d5-f915ebe863bf"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.589530 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d774a24e-5451-4976-a0d5-f915ebe863bf-kube-api-access-h66vg" (OuterVolumeSpecName: "kube-api-access-h66vg") pod "d774a24e-5451-4976-a0d5-f915ebe863bf" (UID: "d774a24e-5451-4976-a0d5-f915ebe863bf"). InnerVolumeSpecName "kube-api-access-h66vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.589650 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-scripts" (OuterVolumeSpecName: "scripts") pod "d774a24e-5451-4976-a0d5-f915ebe863bf" (UID: "d774a24e-5451-4976-a0d5-f915ebe863bf"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.590125 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d774a24e-5451-4976-a0d5-f915ebe863bf" (UID: "d774a24e-5451-4976-a0d5-f915ebe863bf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.590355 4787 scope.go:117] "RemoveContainer" containerID="e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.617855 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-kube-api-access-dbvk7" (OuterVolumeSpecName: "kube-api-access-dbvk7") pod "2ea7d1a3-de18-4283-bd8e-2be80f82eb9a" (UID: "2ea7d1a3-de18-4283-bd8e-2be80f82eb9a"). InnerVolumeSpecName "kube-api-access-dbvk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.618684 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-3101-account-create-update-g8j4t" event={"ID":"2ea7d1a3-de18-4283-bd8e-2be80f82eb9a","Type":"ContainerDied","Data":"32a32fb37176489b55bbf6d661a992c81cf0069e793d28ac2a69ba61336050ea"} Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.618724 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32a32fb37176489b55bbf6d661a992c81cf0069e793d28ac2a69ba61336050ea" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.618779 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-3101-account-create-update-g8j4t" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.633936 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d774a24e-5451-4976-a0d5-f915ebe863bf" (UID: "d774a24e-5451-4976-a0d5-f915ebe863bf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.642043 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75bff842-ba6b-4f84-bff6-054b7292d82e-operator-scripts\") pod \"75bff842-ba6b-4f84-bff6-054b7292d82e\" (UID: \"75bff842-ba6b-4f84-bff6-054b7292d82e\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.642169 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nng29\" (UniqueName: \"kubernetes.io/projected/75bff842-ba6b-4f84-bff6-054b7292d82e-kube-api-access-nng29\") pod \"75bff842-ba6b-4f84-bff6-054b7292d82e\" (UID: \"75bff842-ba6b-4f84-bff6-054b7292d82e\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.642907 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbvk7\" (UniqueName: \"kubernetes.io/projected/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-kube-api-access-dbvk7\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.642929 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5b9s9\" (UniqueName: \"kubernetes.io/projected/244e40a3-b432-4971-bdf5-f070ab9891b0-kube-api-access-5b9s9\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.642942 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.642954 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d774a24e-5451-4976-a0d5-f915ebe863bf-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.642965 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.642977 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.642989 4787 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/d774a24e-5451-4976-a0d5-f915ebe863bf-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.643001 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.643033 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h66vg\" (UniqueName: \"kubernetes.io/projected/d774a24e-5451-4976-a0d5-f915ebe863bf-kube-api-access-h66vg\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.643004 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/75bff842-ba6b-4f84-bff6-054b7292d82e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "75bff842-ba6b-4f84-bff6-054b7292d82e" (UID: "75bff842-ba6b-4f84-bff6-054b7292d82e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.670181 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/75bff842-ba6b-4f84-bff6-054b7292d82e-kube-api-access-nng29" (OuterVolumeSpecName: "kube-api-access-nng29") pod "75bff842-ba6b-4f84-bff6-054b7292d82e" (UID: "75bff842-ba6b-4f84-bff6-054b7292d82e"). InnerVolumeSpecName "kube-api-access-nng29". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.683576 4787 generic.go:334] "Generic (PLEG): container finished" podID="dc14f381-14f6-4719-a0a9-69bfc737fcb5" containerID="4786ee5b7941d5e9659f45efe86cd1ca5e32ea0024e2d05b1cd3908ec5f53202" exitCode=0 Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.684860 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5bcbf94467-l2njc" event={"ID":"dc14f381-14f6-4719-a0a9-69bfc737fcb5","Type":"ContainerDied","Data":"4786ee5b7941d5e9659f45efe86cd1ca5e32ea0024e2d05b1cd3908ec5f53202"} Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.691839 4787 scope.go:117] "RemoveContainer" containerID="39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.704424 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-65b7898d75-fjwnb" event={"ID":"cf9c3742-1e29-4f87-bb53-9a60ab3c14ee","Type":"ContainerStarted","Data":"56bd2bfb43f489879452d3560fdb51baa5b54a1d886acaaf4f578fe1a9884217"} Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.708326 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=19.708299779 podStartE2EDuration="19.708299779s" podCreationTimestamp="2025-12-03 17:37:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:37:26.634234844 +0000 UTC m=+1463.451705803" watchObservedRunningTime="2025-12-03 17:37:26.708299779 +0000 UTC m=+1463.525770738" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.707031 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.728470 4787 generic.go:334] "Generic (PLEG): container finished" podID="9296f967-7bf3-4e42-a867-cce34092dc11" containerID="916e6bd263c38203861c58ce0127628fa60d0530c78a9d5fab6c5aa9394d2f5c" exitCode=1 Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.728569 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7994c4bbc4-2kvbc" event={"ID":"9296f967-7bf3-4e42-a867-cce34092dc11","Type":"ContainerDied","Data":"916e6bd263c38203861c58ce0127628fa60d0530c78a9d5fab6c5aa9394d2f5c"} Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.729425 4787 scope.go:117] "RemoveContainer" containerID="916e6bd263c38203861c58ce0127628fa60d0530c78a9d5fab6c5aa9394d2f5c" Dec 03 17:37:26 crc kubenswrapper[4787]: E1203 17:37:26.729723 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7994c4bbc4-2kvbc_openstack(9296f967-7bf3-4e42-a867-cce34092dc11)\"" pod="openstack/heat-api-7994c4bbc4-2kvbc" podUID="9296f967-7bf3-4e42-a867-cce34092dc11" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.746373 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/75bff842-ba6b-4f84-bff6-054b7292d82e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.746437 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nng29\" (UniqueName: \"kubernetes.io/projected/75bff842-ba6b-4f84-bff6-054b7292d82e-kube-api-access-nng29\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.746942 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.748065 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-hj85k" event={"ID":"6371d53b-b035-4a6e-82b5-c5774ffc8de1","Type":"ContainerDied","Data":"d0f135984c8ebe69810235a15d965cad8f9fa74d63fafb683d170f6bc9322bd4"} Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.748099 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0f135984c8ebe69810235a15d965cad8f9fa74d63fafb683d170f6bc9322bd4" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.748147 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-hj85k" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.751067 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data" (OuterVolumeSpecName: "config-data") pod "d774a24e-5451-4976-a0d5-f915ebe863bf" (UID: "d774a24e-5451-4976-a0d5-f915ebe863bf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.766349 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-65b7898d75-fjwnb" podStartSLOduration=18.76632647 podStartE2EDuration="18.76632647s" podCreationTimestamp="2025-12-03 17:37:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:37:26.740348666 +0000 UTC m=+1463.557819625" watchObservedRunningTime="2025-12-03 17:37:26.76632647 +0000 UTC m=+1463.583797429" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.766645 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.777207 4787 scope.go:117] "RemoveContainer" containerID="e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183" Dec 03 17:37:26 crc kubenswrapper[4787]: E1203 17:37:26.781801 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183\": container with ID starting with e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183 not found: ID does not exist" containerID="e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.781853 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183"} err="failed to get container status \"e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183\": rpc error: code = NotFound desc = could not find container \"e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183\": container with ID starting with e1cc5d6a78e607edab78ad2ccc2fa1b4da080738ec7382600573cdd1acddc183 not found: ID does not exist" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.781877 4787 scope.go:117] "RemoveContainer" containerID="39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5" Dec 03 17:37:26 crc kubenswrapper[4787]: E1203 17:37:26.791258 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5\": container with ID starting with 39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5 not found: ID does not exist" containerID="39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.791320 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5"} err="failed to get container status \"39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5\": rpc error: code = NotFound desc = could not find container \"39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5\": container with ID starting with 39792f7a67b350dc1f7e598e3124823b6fcdf2a1899751f0bfec2d38090225d5 not found: ID does not exist" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.791351 4787 scope.go:117] "RemoveContainer" containerID="23f7168ce7eebdba833bc2c4cb8e6e1b44620b162f312222dbde84b090254536" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.847672 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data\") pod \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.847762 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-combined-ca-bundle\") pod \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.847839 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data-custom\") pod \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.848010 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5qrs\" (UniqueName: \"kubernetes.io/projected/dc14f381-14f6-4719-a0a9-69bfc737fcb5-kube-api-access-l5qrs\") pod \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\" (UID: \"dc14f381-14f6-4719-a0a9-69bfc737fcb5\") " Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.848706 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d774a24e-5451-4976-a0d5-f915ebe863bf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.861156 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "dc14f381-14f6-4719-a0a9-69bfc737fcb5" (UID: "dc14f381-14f6-4719-a0a9-69bfc737fcb5"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.865871 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc14f381-14f6-4719-a0a9-69bfc737fcb5-kube-api-access-l5qrs" (OuterVolumeSpecName: "kube-api-access-l5qrs") pod "dc14f381-14f6-4719-a0a9-69bfc737fcb5" (UID: "dc14f381-14f6-4719-a0a9-69bfc737fcb5"). InnerVolumeSpecName "kube-api-access-l5qrs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.895194 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dc14f381-14f6-4719-a0a9-69bfc737fcb5" (UID: "dc14f381-14f6-4719-a0a9-69bfc737fcb5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.930196 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data" (OuterVolumeSpecName: "config-data") pod "dc14f381-14f6-4719-a0a9-69bfc737fcb5" (UID: "dc14f381-14f6-4719-a0a9-69bfc737fcb5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.951556 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.951606 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.951619 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5qrs\" (UniqueName: \"kubernetes.io/projected/dc14f381-14f6-4719-a0a9-69bfc737fcb5-kube-api-access-l5qrs\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:26 crc kubenswrapper[4787]: I1203 17:37:26.951633 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dc14f381-14f6-4719-a0a9-69bfc737fcb5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.171133 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.215419 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.286132 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:37:27 crc kubenswrapper[4787]: E1203 17:37:27.286825 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d774a24e-5451-4976-a0d5-f915ebe863bf" containerName="cinder-api-log" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.286839 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d774a24e-5451-4976-a0d5-f915ebe863bf" containerName="cinder-api-log" Dec 03 17:37:27 crc kubenswrapper[4787]: E1203 17:37:27.286851 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2ea7d1a3-de18-4283-bd8e-2be80f82eb9a" containerName="mariadb-account-create-update" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.286858 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2ea7d1a3-de18-4283-bd8e-2be80f82eb9a" containerName="mariadb-account-create-update" Dec 03 17:37:27 crc kubenswrapper[4787]: E1203 17:37:27.286871 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="244e40a3-b432-4971-bdf5-f070ab9891b0" containerName="mariadb-database-create" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.286877 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="244e40a3-b432-4971-bdf5-f070ab9891b0" containerName="mariadb-database-create" Dec 03 17:37:27 crc kubenswrapper[4787]: E1203 17:37:27.286886 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc14f381-14f6-4719-a0a9-69bfc737fcb5" containerName="heat-cfnapi" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.286892 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc14f381-14f6-4719-a0a9-69bfc737fcb5" containerName="heat-cfnapi" Dec 03 17:37:27 crc kubenswrapper[4787]: E1203 17:37:27.286907 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="75bff842-ba6b-4f84-bff6-054b7292d82e" containerName="mariadb-account-create-update" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.286912 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="75bff842-ba6b-4f84-bff6-054b7292d82e" containerName="mariadb-account-create-update" Dec 03 17:37:27 crc kubenswrapper[4787]: E1203 17:37:27.286925 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d774a24e-5451-4976-a0d5-f915ebe863bf" containerName="cinder-api" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.286930 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d774a24e-5451-4976-a0d5-f915ebe863bf" containerName="cinder-api" Dec 03 17:37:27 crc kubenswrapper[4787]: E1203 17:37:27.286937 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="39227401-0c97-4fd7-8510-615e22ca73d9" containerName="mariadb-account-create-update" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.286942 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="39227401-0c97-4fd7-8510-615e22ca73d9" containerName="mariadb-account-create-update" Dec 03 17:37:27 crc kubenswrapper[4787]: E1203 17:37:27.286954 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6371d53b-b035-4a6e-82b5-c5774ffc8de1" containerName="mariadb-database-create" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.286959 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6371d53b-b035-4a6e-82b5-c5774ffc8de1" containerName="mariadb-database-create" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.287191 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="75bff842-ba6b-4f84-bff6-054b7292d82e" containerName="mariadb-account-create-update" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.287203 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="39227401-0c97-4fd7-8510-615e22ca73d9" containerName="mariadb-account-create-update" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.287218 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d774a24e-5451-4976-a0d5-f915ebe863bf" containerName="cinder-api" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.287229 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="244e40a3-b432-4971-bdf5-f070ab9891b0" containerName="mariadb-database-create" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.287241 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2ea7d1a3-de18-4283-bd8e-2be80f82eb9a" containerName="mariadb-account-create-update" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.287250 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d774a24e-5451-4976-a0d5-f915ebe863bf" containerName="cinder-api-log" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.287256 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6371d53b-b035-4a6e-82b5-c5774ffc8de1" containerName="mariadb-database-create" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.287270 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc14f381-14f6-4719-a0a9-69bfc737fcb5" containerName="heat-cfnapi" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.288372 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.301333 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.301500 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.301600 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.346959 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.364284 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-config-data-custom\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.364336 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-config-data\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.364401 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.364430 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.364446 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-scripts\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.364462 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-logs\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.364495 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52n2v\" (UniqueName: \"kubernetes.io/projected/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-kube-api-access-52n2v\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.364537 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.364561 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.470657 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-config-data-custom\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.471498 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-config-data\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.471894 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.471954 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.471973 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-scripts\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.471993 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-logs\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.472060 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52n2v\" (UniqueName: \"kubernetes.io/projected/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-kube-api-access-52n2v\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.472189 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.472232 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.473148 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.473439 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-logs\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.477694 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-public-tls-certs\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.478547 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.481416 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-config-data\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.481906 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-config-data-custom\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.482303 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.489531 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-scripts\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.492847 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52n2v\" (UniqueName: \"kubernetes.io/projected/e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233-kube-api-access-52n2v\") pod \"cinder-api-0\" (UID: \"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233\") " pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.493996 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6dhs9" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.508369 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.573926 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a23593c-7762-4639-8b2f-2cdd49b0c728-operator-scripts\") pod \"1a23593c-7762-4639-8b2f-2cdd49b0c728\" (UID: \"1a23593c-7762-4639-8b2f-2cdd49b0c728\") " Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.575167 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-combined-ca-bundle\") pod \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.575222 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data\") pod \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.575319 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vnjw7\" (UniqueName: \"kubernetes.io/projected/455e1ae0-75b9-4c80-9580-7b2ba4c72483-kube-api-access-vnjw7\") pod \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.575327 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1a23593c-7762-4639-8b2f-2cdd49b0c728-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1a23593c-7762-4639-8b2f-2cdd49b0c728" (UID: "1a23593c-7762-4639-8b2f-2cdd49b0c728"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.575344 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data-custom\") pod \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\" (UID: \"455e1ae0-75b9-4c80-9580-7b2ba4c72483\") " Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.575391 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bprnc\" (UniqueName: \"kubernetes.io/projected/1a23593c-7762-4639-8b2f-2cdd49b0c728-kube-api-access-bprnc\") pod \"1a23593c-7762-4639-8b2f-2cdd49b0c728\" (UID: \"1a23593c-7762-4639-8b2f-2cdd49b0c728\") " Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.576102 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1a23593c-7762-4639-8b2f-2cdd49b0c728-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.592089 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "455e1ae0-75b9-4c80-9580-7b2ba4c72483" (UID: "455e1ae0-75b9-4c80-9580-7b2ba4c72483"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.593705 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/455e1ae0-75b9-4c80-9580-7b2ba4c72483-kube-api-access-vnjw7" (OuterVolumeSpecName: "kube-api-access-vnjw7") pod "455e1ae0-75b9-4c80-9580-7b2ba4c72483" (UID: "455e1ae0-75b9-4c80-9580-7b2ba4c72483"). InnerVolumeSpecName "kube-api-access-vnjw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.593810 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1a23593c-7762-4639-8b2f-2cdd49b0c728-kube-api-access-bprnc" (OuterVolumeSpecName: "kube-api-access-bprnc") pod "1a23593c-7762-4639-8b2f-2cdd49b0c728" (UID: "1a23593c-7762-4639-8b2f-2cdd49b0c728"). InnerVolumeSpecName "kube-api-access-bprnc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.606491 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.624471 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "455e1ae0-75b9-4c80-9580-7b2ba4c72483" (UID: "455e1ae0-75b9-4c80-9580-7b2ba4c72483"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.658128 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data" (OuterVolumeSpecName: "config-data") pod "455e1ae0-75b9-4c80-9580-7b2ba4c72483" (UID: "455e1ae0-75b9-4c80-9580-7b2ba4c72483"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.679178 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vnjw7\" (UniqueName: \"kubernetes.io/projected/455e1ae0-75b9-4c80-9580-7b2ba4c72483-kube-api-access-vnjw7\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.679214 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.679223 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bprnc\" (UniqueName: \"kubernetes.io/projected/1a23593c-7762-4639-8b2f-2cdd49b0c728-kube-api-access-bprnc\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.679232 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.679242 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/455e1ae0-75b9-4c80-9580-7b2ba4c72483-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.815600 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d774a24e-5451-4976-a0d5-f915ebe863bf" path="/var/lib/kubelet/pods/d774a24e-5451-4976-a0d5-f915ebe863bf/volumes" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.838681 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-lcm5g" event={"ID":"244e40a3-b432-4971-bdf5-f070ab9891b0","Type":"ContainerDied","Data":"5b08eb4f7625d13fe465489f5edd5af57e7f41abbcd4daa9f406bca990cf4a02"} Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.838726 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b08eb4f7625d13fe465489f5edd5af57e7f41abbcd4daa9f406bca990cf4a02" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.838817 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-lcm5g" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.855918 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-6dhs9" event={"ID":"1a23593c-7762-4639-8b2f-2cdd49b0c728","Type":"ContainerDied","Data":"e9d7997a08de7b48222229340121e1186628293d7f3db99b553e932840f8606f"} Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.855962 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9d7997a08de7b48222229340121e1186628293d7f3db99b553e932840f8606f" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.856045 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-6dhs9" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.861924 4787 scope.go:117] "RemoveContainer" containerID="7e594040c2aee27dd4dac76f8492ed0427c1456eebe85855a32e5f68443f301a" Dec 03 17:37:27 crc kubenswrapper[4787]: E1203 17:37:27.864577 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-64cb8b689b-d4pmz_openstack(8f12c699-048b-4eee-bc7c-e3ad916656ed)\"" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" podUID="8f12c699-048b-4eee-bc7c-e3ad916656ed" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.867895 4787 scope.go:117] "RemoveContainer" containerID="916e6bd263c38203861c58ce0127628fa60d0530c78a9d5fab6c5aa9394d2f5c" Dec 03 17:37:27 crc kubenswrapper[4787]: E1203 17:37:27.868194 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7994c4bbc4-2kvbc_openstack(9296f967-7bf3-4e42-a867-cce34092dc11)\"" pod="openstack/heat-api-7994c4bbc4-2kvbc" podUID="9296f967-7bf3-4e42-a867-cce34092dc11" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.869502 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-91eb-account-create-update-59l2z" event={"ID":"75bff842-ba6b-4f84-bff6-054b7292d82e","Type":"ContainerDied","Data":"60aec9976e5ab8829cde63b74025abc8da5dcf4a7e9952444eaf0709c0323e9a"} Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.869598 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="60aec9976e5ab8829cde63b74025abc8da5dcf4a7e9952444eaf0709c0323e9a" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.869722 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-91eb-account-create-update-59l2z" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.918072 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"956e5a96-03a5-40ca-a9ce-fbdc8631edb3","Type":"ContainerStarted","Data":"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af"} Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.922087 4787 generic.go:334] "Generic (PLEG): container finished" podID="455e1ae0-75b9-4c80-9580-7b2ba4c72483" containerID="030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5" exitCode=0 Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.922259 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-75d5f9cb88-48rbj" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.922346 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-75d5f9cb88-48rbj" event={"ID":"455e1ae0-75b9-4c80-9580-7b2ba4c72483","Type":"ContainerDied","Data":"030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5"} Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.922450 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-75d5f9cb88-48rbj" event={"ID":"455e1ae0-75b9-4c80-9580-7b2ba4c72483","Type":"ContainerDied","Data":"ff728de928f4390ce704b9690b0fc0885b42104a3518c2fcdfc2c8b6389a6254"} Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.922714 4787 scope.go:117] "RemoveContainer" containerID="030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.955878 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5bcbf94467-l2njc" Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.956451 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5bcbf94467-l2njc" event={"ID":"dc14f381-14f6-4719-a0a9-69bfc737fcb5","Type":"ContainerDied","Data":"e51c8d5dcb395a26e64db3d2bbfe9fba858f9d1312e51c473fc997b5ea75f108"} Dec 03 17:37:27 crc kubenswrapper[4787]: I1203 17:37:27.956934 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.001168 4787 scope.go:117] "RemoveContainer" containerID="030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5" Dec 03 17:37:28 crc kubenswrapper[4787]: E1203 17:37:28.002371 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5\": container with ID starting with 030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5 not found: ID does not exist" containerID="030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5" Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.002396 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5"} err="failed to get container status \"030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5\": rpc error: code = NotFound desc = could not find container \"030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5\": container with ID starting with 030af5b1b797f19631ac09f65a0df5b3018169baa1bf692165e1b7022b516ae5 not found: ID does not exist" Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.002417 4787 scope.go:117] "RemoveContainer" containerID="4786ee5b7941d5e9659f45efe86cd1ca5e32ea0024e2d05b1cd3908ec5f53202" Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.117010 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.219693 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5bcbf94467-l2njc"] Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.237393 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-5bcbf94467-l2njc"] Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.251097 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-75d5f9cb88-48rbj"] Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.276981 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-75d5f9cb88-48rbj"] Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.525774 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 03 17:37:28 crc kubenswrapper[4787]: W1203 17:37:28.528903 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode4a9ddfc_3167_4d7c_bd2a_6ae7a7724233.slice/crio-e60b85ba806500f004a1fa49422c5fab33cf39f45b11bb3ea59212925eab97af WatchSource:0}: Error finding container e60b85ba806500f004a1fa49422c5fab33cf39f45b11bb3ea59212925eab97af: Status 404 returned error can't find the container with id e60b85ba806500f004a1fa49422c5fab33cf39f45b11bb3ea59212925eab97af Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.965659 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233","Type":"ContainerStarted","Data":"e60b85ba806500f004a1fa49422c5fab33cf39f45b11bb3ea59212925eab97af"} Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.982664 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"956e5a96-03a5-40ca-a9ce-fbdc8631edb3","Type":"ContainerStarted","Data":"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1"} Dec 03 17:37:28 crc kubenswrapper[4787]: I1203 17:37:28.983641 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:37:29 crc kubenswrapper[4787]: I1203 17:37:29.021059 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.4463031060000002 podStartE2EDuration="8.021041827s" podCreationTimestamp="2025-12-03 17:37:21 +0000 UTC" firstStartedPulling="2025-12-03 17:37:23.132261383 +0000 UTC m=+1459.949732342" lastFinishedPulling="2025-12-03 17:37:27.707000104 +0000 UTC m=+1464.524471063" observedRunningTime="2025-12-03 17:37:29.019267999 +0000 UTC m=+1465.836738978" watchObservedRunningTime="2025-12-03 17:37:29.021041827 +0000 UTC m=+1465.838512786" Dec 03 17:37:29 crc kubenswrapper[4787]: I1203 17:37:29.146768 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:29 crc kubenswrapper[4787]: I1203 17:37:29.780160 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="455e1ae0-75b9-4c80-9580-7b2ba4c72483" path="/var/lib/kubelet/pods/455e1ae0-75b9-4c80-9580-7b2ba4c72483/volumes" Dec 03 17:37:29 crc kubenswrapper[4787]: I1203 17:37:29.780929 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc14f381-14f6-4719-a0a9-69bfc737fcb5" path="/var/lib/kubelet/pods/dc14f381-14f6-4719-a0a9-69bfc737fcb5/volumes" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.000609 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233","Type":"ContainerStarted","Data":"6bb5b4f27c0b1128789308c9704c7a22577b04aff784fe04c6f2903ad497aeeb"} Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.013379 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-65b7898d75-fjwnb" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.404536 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.462547 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.462602 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.463111 4787 scope.go:117] "RemoveContainer" containerID="916e6bd263c38203861c58ce0127628fa60d0530c78a9d5fab6c5aa9394d2f5c" Dec 03 17:37:30 crc kubenswrapper[4787]: E1203 17:37:30.463458 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-api pod=heat-api-7994c4bbc4-2kvbc_openstack(9296f967-7bf3-4e42-a867-cce34092dc11)\"" pod="openstack/heat-api-7994c4bbc4-2kvbc" podUID="9296f967-7bf3-4e42-a867-cce34092dc11" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.478898 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.478939 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.479742 4787 scope.go:117] "RemoveContainer" containerID="7e594040c2aee27dd4dac76f8492ed0427c1456eebe85855a32e5f68443f301a" Dec 03 17:37:30 crc kubenswrapper[4787]: E1203 17:37:30.480009 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-cfnapi\" with CrashLoopBackOff: \"back-off 10s restarting failed container=heat-cfnapi pod=heat-cfnapi-64cb8b689b-d4pmz_openstack(8f12c699-048b-4eee-bc7c-e3ad916656ed)\"" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" podUID="8f12c699-048b-4eee-bc7c-e3ad916656ed" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.729960 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l2pkv"] Dec 03 17:37:30 crc kubenswrapper[4787]: E1203 17:37:30.730515 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="455e1ae0-75b9-4c80-9580-7b2ba4c72483" containerName="heat-api" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.730538 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="455e1ae0-75b9-4c80-9580-7b2ba4c72483" containerName="heat-api" Dec 03 17:37:30 crc kubenswrapper[4787]: E1203 17:37:30.730553 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1a23593c-7762-4639-8b2f-2cdd49b0c728" containerName="mariadb-database-create" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.730559 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1a23593c-7762-4639-8b2f-2cdd49b0c728" containerName="mariadb-database-create" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.730788 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1a23593c-7762-4639-8b2f-2cdd49b0c728" containerName="mariadb-database-create" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.730825 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="455e1ae0-75b9-4c80-9580-7b2ba4c72483" containerName="heat-api" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.731680 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.738148 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-669qv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.738216 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.738245 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.749584 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l2pkv"] Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.870588 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-config-data\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.871004 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.871197 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-scripts\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.871276 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nfqm\" (UniqueName: \"kubernetes.io/projected/131d89aa-86f5-4ddb-a388-f2a938319ce2-kube-api-access-7nfqm\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.973656 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-scripts\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.973714 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nfqm\" (UniqueName: \"kubernetes.io/projected/131d89aa-86f5-4ddb-a388-f2a938319ce2-kube-api-access-7nfqm\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.973842 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-config-data\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.973883 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.979272 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.979277 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-scripts\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.981267 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-config-data\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:30 crc kubenswrapper[4787]: I1203 17:37:30.993005 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nfqm\" (UniqueName: \"kubernetes.io/projected/131d89aa-86f5-4ddb-a388-f2a938319ce2-kube-api-access-7nfqm\") pod \"nova-cell0-conductor-db-sync-l2pkv\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:31 crc kubenswrapper[4787]: I1203 17:37:31.014569 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233","Type":"ContainerStarted","Data":"b37c4a5718ee3393e450af16bf04f5563f16ce5fb170478d388745efb42a7859"} Dec 03 17:37:31 crc kubenswrapper[4787]: I1203 17:37:31.015272 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 03 17:37:31 crc kubenswrapper[4787]: I1203 17:37:31.047996 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.047974856 podStartE2EDuration="4.047974856s" podCreationTimestamp="2025-12-03 17:37:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:37:31.03704312 +0000 UTC m=+1467.854514089" watchObservedRunningTime="2025-12-03 17:37:31.047974856 +0000 UTC m=+1467.865445815" Dec 03 17:37:31 crc kubenswrapper[4787]: I1203 17:37:31.077501 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:37:31 crc kubenswrapper[4787]: I1203 17:37:31.581364 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l2pkv"] Dec 03 17:37:31 crc kubenswrapper[4787]: W1203 17:37:31.581926 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod131d89aa_86f5_4ddb_a388_f2a938319ce2.slice/crio-8c5610c461e1449247e5339eb15bc2ce56923144754d4a53c718ab3759ceeb62 WatchSource:0}: Error finding container 8c5610c461e1449247e5339eb15bc2ce56923144754d4a53c718ab3759ceeb62: Status 404 returned error can't find the container with id 8c5610c461e1449247e5339eb15bc2ce56923144754d4a53c718ab3759ceeb62 Dec 03 17:37:31 crc kubenswrapper[4787]: I1203 17:37:31.820444 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:37:31 crc kubenswrapper[4787]: I1203 17:37:31.899993 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6slhs"] Dec 03 17:37:31 crc kubenswrapper[4787]: I1203 17:37:31.905770 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" podUID="88c9db22-be72-4ff3-a4bf-361d634922b0" containerName="dnsmasq-dns" containerID="cri-o://73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc" gracePeriod=10 Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.033326 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l2pkv" event={"ID":"131d89aa-86f5-4ddb-a388-f2a938319ce2","Type":"ContainerStarted","Data":"8c5610c461e1449247e5339eb15bc2ce56923144754d4a53c718ab3759ceeb62"} Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.034079 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="ceilometer-central-agent" containerID="cri-o://82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a" gracePeriod=30 Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.034222 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="proxy-httpd" containerID="cri-o://93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1" gracePeriod=30 Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.034280 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="sg-core" containerID="cri-o://daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af" gracePeriod=30 Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.034324 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="ceilometer-notification-agent" containerID="cri-o://4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568" gracePeriod=30 Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.530709 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.628867 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-config\") pod \"88c9db22-be72-4ff3-a4bf-361d634922b0\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.628952 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-nb\") pod \"88c9db22-be72-4ff3-a4bf-361d634922b0\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.629004 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-sb\") pod \"88c9db22-be72-4ff3-a4bf-361d634922b0\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.629057 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-swift-storage-0\") pod \"88c9db22-be72-4ff3-a4bf-361d634922b0\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.629340 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-svc\") pod \"88c9db22-be72-4ff3-a4bf-361d634922b0\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.629374 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fdnm4\" (UniqueName: \"kubernetes.io/projected/88c9db22-be72-4ff3-a4bf-361d634922b0-kube-api-access-fdnm4\") pod \"88c9db22-be72-4ff3-a4bf-361d634922b0\" (UID: \"88c9db22-be72-4ff3-a4bf-361d634922b0\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.643673 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88c9db22-be72-4ff3-a4bf-361d634922b0-kube-api-access-fdnm4" (OuterVolumeSpecName: "kube-api-access-fdnm4") pod "88c9db22-be72-4ff3-a4bf-361d634922b0" (UID: "88c9db22-be72-4ff3-a4bf-361d634922b0"). InnerVolumeSpecName "kube-api-access-fdnm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.701617 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-config" (OuterVolumeSpecName: "config") pod "88c9db22-be72-4ff3-a4bf-361d634922b0" (UID: "88c9db22-be72-4ff3-a4bf-361d634922b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.718562 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "88c9db22-be72-4ff3-a4bf-361d634922b0" (UID: "88c9db22-be72-4ff3-a4bf-361d634922b0"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.719076 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "88c9db22-be72-4ff3-a4bf-361d634922b0" (UID: "88c9db22-be72-4ff3-a4bf-361d634922b0"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.719855 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "88c9db22-be72-4ff3-a4bf-361d634922b0" (UID: "88c9db22-be72-4ff3-a4bf-361d634922b0"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.731758 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.731801 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fdnm4\" (UniqueName: \"kubernetes.io/projected/88c9db22-be72-4ff3-a4bf-361d634922b0-kube-api-access-fdnm4\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.731816 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.731830 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.731841 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.737477 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "88c9db22-be72-4ff3-a4bf-361d634922b0" (UID: "88c9db22-be72-4ff3-a4bf-361d634922b0"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.833410 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/88c9db22-be72-4ff3-a4bf-361d634922b0-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.870607 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.934908 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-log-httpd\") pod \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.935101 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-run-httpd\") pod \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.935145 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-sg-core-conf-yaml\") pod \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.935275 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-combined-ca-bundle\") pod \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.935326 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-config-data\") pod \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.935393 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-scripts\") pod \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.935435 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g2c5d\" (UniqueName: \"kubernetes.io/projected/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-kube-api-access-g2c5d\") pod \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\" (UID: \"956e5a96-03a5-40ca-a9ce-fbdc8631edb3\") " Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.935434 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "956e5a96-03a5-40ca-a9ce-fbdc8631edb3" (UID: "956e5a96-03a5-40ca-a9ce-fbdc8631edb3"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.935559 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "956e5a96-03a5-40ca-a9ce-fbdc8631edb3" (UID: "956e5a96-03a5-40ca-a9ce-fbdc8631edb3"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.936130 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.936151 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.939597 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-scripts" (OuterVolumeSpecName: "scripts") pod "956e5a96-03a5-40ca-a9ce-fbdc8631edb3" (UID: "956e5a96-03a5-40ca-a9ce-fbdc8631edb3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.943355 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-kube-api-access-g2c5d" (OuterVolumeSpecName: "kube-api-access-g2c5d") pod "956e5a96-03a5-40ca-a9ce-fbdc8631edb3" (UID: "956e5a96-03a5-40ca-a9ce-fbdc8631edb3"). InnerVolumeSpecName "kube-api-access-g2c5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:32 crc kubenswrapper[4787]: I1203 17:37:32.976520 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "956e5a96-03a5-40ca-a9ce-fbdc8631edb3" (UID: "956e5a96-03a5-40ca-a9ce-fbdc8631edb3"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.038178 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.038209 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.038220 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g2c5d\" (UniqueName: \"kubernetes.io/projected/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-kube-api-access-g2c5d\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.061360 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "956e5a96-03a5-40ca-a9ce-fbdc8631edb3" (UID: "956e5a96-03a5-40ca-a9ce-fbdc8631edb3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068184 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-config-data" (OuterVolumeSpecName: "config-data") pod "956e5a96-03a5-40ca-a9ce-fbdc8631edb3" (UID: "956e5a96-03a5-40ca-a9ce-fbdc8631edb3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068364 4787 generic.go:334] "Generic (PLEG): container finished" podID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerID="93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1" exitCode=0 Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068396 4787 generic.go:334] "Generic (PLEG): container finished" podID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerID="daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af" exitCode=2 Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068404 4787 generic.go:334] "Generic (PLEG): container finished" podID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerID="4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568" exitCode=0 Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068412 4787 generic.go:334] "Generic (PLEG): container finished" podID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerID="82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a" exitCode=0 Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068500 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"956e5a96-03a5-40ca-a9ce-fbdc8631edb3","Type":"ContainerDied","Data":"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1"} Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068532 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"956e5a96-03a5-40ca-a9ce-fbdc8631edb3","Type":"ContainerDied","Data":"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af"} Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068545 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"956e5a96-03a5-40ca-a9ce-fbdc8631edb3","Type":"ContainerDied","Data":"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568"} Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068557 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"956e5a96-03a5-40ca-a9ce-fbdc8631edb3","Type":"ContainerDied","Data":"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a"} Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068567 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"956e5a96-03a5-40ca-a9ce-fbdc8631edb3","Type":"ContainerDied","Data":"7700d42651bc8bd214ea8f1fb2ff339b3a67dca1566dc6dc7ab5a224ee680d2c"} Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068471 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.068584 4787 scope.go:117] "RemoveContainer" containerID="93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.080388 4787 generic.go:334] "Generic (PLEG): container finished" podID="88c9db22-be72-4ff3-a4bf-361d634922b0" containerID="73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc" exitCode=0 Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.081801 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.082150 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" event={"ID":"88c9db22-be72-4ff3-a4bf-361d634922b0","Type":"ContainerDied","Data":"73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc"} Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.082199 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5c9776ccc5-6slhs" event={"ID":"88c9db22-be72-4ff3-a4bf-361d634922b0","Type":"ContainerDied","Data":"a4aa3563276ddeb0e749c90c679b01ee2cb91f766bd987d726b6497f16fb8b6b"} Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.140582 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.140622 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/956e5a96-03a5-40ca-a9ce-fbdc8631edb3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.240959 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.244376 4787 scope.go:117] "RemoveContainer" containerID="daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.279095 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.288087 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6slhs"] Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.294781 4787 scope.go:117] "RemoveContainer" containerID="4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.301582 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5c9776ccc5-6slhs"] Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.316857 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.317420 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c9db22-be72-4ff3-a4bf-361d634922b0" containerName="init" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.317435 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c9db22-be72-4ff3-a4bf-361d634922b0" containerName="init" Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.317469 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88c9db22-be72-4ff3-a4bf-361d634922b0" containerName="dnsmasq-dns" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.317477 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="88c9db22-be72-4ff3-a4bf-361d634922b0" containerName="dnsmasq-dns" Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.317529 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="ceilometer-central-agent" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.317539 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="ceilometer-central-agent" Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.317554 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="ceilometer-notification-agent" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.317561 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="ceilometer-notification-agent" Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.317573 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="sg-core" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.317580 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="sg-core" Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.317591 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="proxy-httpd" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.317598 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="proxy-httpd" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.317838 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="88c9db22-be72-4ff3-a4bf-361d634922b0" containerName="dnsmasq-dns" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.317857 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="sg-core" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.317869 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="ceilometer-notification-agent" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.317882 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="ceilometer-central-agent" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.317901 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" containerName="proxy-httpd" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.320279 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.324315 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.324520 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.335162 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.336654 4787 scope.go:117] "RemoveContainer" containerID="82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.369659 4787 scope.go:117] "RemoveContainer" containerID="93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1" Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.373503 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1\": container with ID starting with 93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1 not found: ID does not exist" containerID="93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.373552 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1"} err="failed to get container status \"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1\": rpc error: code = NotFound desc = could not find container \"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1\": container with ID starting with 93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1 not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.373585 4787 scope.go:117] "RemoveContainer" containerID="daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af" Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.374167 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af\": container with ID starting with daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af not found: ID does not exist" containerID="daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.374202 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af"} err="failed to get container status \"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af\": rpc error: code = NotFound desc = could not find container \"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af\": container with ID starting with daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.374219 4787 scope.go:117] "RemoveContainer" containerID="4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568" Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.374508 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568\": container with ID starting with 4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568 not found: ID does not exist" containerID="4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.374538 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568"} err="failed to get container status \"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568\": rpc error: code = NotFound desc = could not find container \"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568\": container with ID starting with 4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568 not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.374556 4787 scope.go:117] "RemoveContainer" containerID="82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a" Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.374774 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a\": container with ID starting with 82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a not found: ID does not exist" containerID="82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.374799 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a"} err="failed to get container status \"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a\": rpc error: code = NotFound desc = could not find container \"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a\": container with ID starting with 82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.374816 4787 scope.go:117] "RemoveContainer" containerID="93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.375114 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1"} err="failed to get container status \"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1\": rpc error: code = NotFound desc = could not find container \"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1\": container with ID starting with 93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1 not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.375141 4787 scope.go:117] "RemoveContainer" containerID="daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.375338 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af"} err="failed to get container status \"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af\": rpc error: code = NotFound desc = could not find container \"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af\": container with ID starting with daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.375413 4787 scope.go:117] "RemoveContainer" containerID="4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.375622 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568"} err="failed to get container status \"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568\": rpc error: code = NotFound desc = could not find container \"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568\": container with ID starting with 4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568 not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.375649 4787 scope.go:117] "RemoveContainer" containerID="82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.375879 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a"} err="failed to get container status \"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a\": rpc error: code = NotFound desc = could not find container \"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a\": container with ID starting with 82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.375896 4787 scope.go:117] "RemoveContainer" containerID="93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.376108 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1"} err="failed to get container status \"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1\": rpc error: code = NotFound desc = could not find container \"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1\": container with ID starting with 93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1 not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.376125 4787 scope.go:117] "RemoveContainer" containerID="daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.376353 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af"} err="failed to get container status \"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af\": rpc error: code = NotFound desc = could not find container \"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af\": container with ID starting with daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.376380 4787 scope.go:117] "RemoveContainer" containerID="4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.376671 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568"} err="failed to get container status \"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568\": rpc error: code = NotFound desc = could not find container \"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568\": container with ID starting with 4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568 not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.376921 4787 scope.go:117] "RemoveContainer" containerID="82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.377397 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a"} err="failed to get container status \"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a\": rpc error: code = NotFound desc = could not find container \"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a\": container with ID starting with 82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.377437 4787 scope.go:117] "RemoveContainer" containerID="93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.379554 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1"} err="failed to get container status \"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1\": rpc error: code = NotFound desc = could not find container \"93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1\": container with ID starting with 93e38274bae13da0d829a8711925bb6216abd7c03611505f6e955072f828b9b1 not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.379581 4787 scope.go:117] "RemoveContainer" containerID="daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.380261 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af"} err="failed to get container status \"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af\": rpc error: code = NotFound desc = could not find container \"daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af\": container with ID starting with daeebd2f59279ef9d7c547889d05f6fa26fc90d0073875a57bbceb9b1f36e0af not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.380316 4787 scope.go:117] "RemoveContainer" containerID="4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.380653 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568"} err="failed to get container status \"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568\": rpc error: code = NotFound desc = could not find container \"4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568\": container with ID starting with 4fd589fc0446b8b5c6ba3ee62fedfa09b60d4dcd1ca914c0ea1433b7274f8568 not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.380705 4787 scope.go:117] "RemoveContainer" containerID="82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.381046 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a"} err="failed to get container status \"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a\": rpc error: code = NotFound desc = could not find container \"82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a\": container with ID starting with 82644d7bc58785820c25797b052843ce3fa98a99c61a515de0ebd548a7e2a02a not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.381074 4787 scope.go:117] "RemoveContainer" containerID="73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.389080 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.417248 4787 scope.go:117] "RemoveContainer" containerID="84bd3b8ce3a556bd11579c6ad475079452d283de51169129f3b775d5cb11c38f" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.451335 4787 scope.go:117] "RemoveContainer" containerID="73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.452266 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zrgws\" (UniqueName: \"kubernetes.io/projected/e2c6617d-aba8-4b39-94b1-c912a95aa352-kube-api-access-zrgws\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.452347 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-log-httpd\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.452378 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.452411 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-scripts\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.452522 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-run-httpd\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.452583 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.452613 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-config-data\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.457297 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc\": container with ID starting with 73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc not found: ID does not exist" containerID="73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.457353 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc"} err="failed to get container status \"73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc\": rpc error: code = NotFound desc = could not find container \"73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc\": container with ID starting with 73a4d5d8e3dfa5ec4b03f10bca82ab92c8b930fd77aeac11ddc2fc4ed07471bc not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.457382 4787 scope.go:117] "RemoveContainer" containerID="84bd3b8ce3a556bd11579c6ad475079452d283de51169129f3b775d5cb11c38f" Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.457856 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84bd3b8ce3a556bd11579c6ad475079452d283de51169129f3b775d5cb11c38f\": container with ID starting with 84bd3b8ce3a556bd11579c6ad475079452d283de51169129f3b775d5cb11c38f not found: ID does not exist" containerID="84bd3b8ce3a556bd11579c6ad475079452d283de51169129f3b775d5cb11c38f" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.457875 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84bd3b8ce3a556bd11579c6ad475079452d283de51169129f3b775d5cb11c38f"} err="failed to get container status \"84bd3b8ce3a556bd11579c6ad475079452d283de51169129f3b775d5cb11c38f\": rpc error: code = NotFound desc = could not find container \"84bd3b8ce3a556bd11579c6ad475079452d283de51169129f3b775d5cb11c38f\": container with ID starting with 84bd3b8ce3a556bd11579c6ad475079452d283de51169129f3b775d5cb11c38f not found: ID does not exist" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.464731 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:33 crc kubenswrapper[4787]: E1203 17:37:33.475393 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-zrgws log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="e2c6617d-aba8-4b39-94b1-c912a95aa352" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.554129 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.554208 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-config-data\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.554250 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zrgws\" (UniqueName: \"kubernetes.io/projected/e2c6617d-aba8-4b39-94b1-c912a95aa352-kube-api-access-zrgws\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.554299 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-log-httpd\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.554471 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.554513 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-scripts\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.554629 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-run-httpd\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.555144 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-run-httpd\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.556074 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-log-httpd\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.561651 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.562213 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-config-data\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.562677 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.562819 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-scripts\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.581711 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zrgws\" (UniqueName: \"kubernetes.io/projected/e2c6617d-aba8-4b39-94b1-c912a95aa352-kube-api-access-zrgws\") pod \"ceilometer-0\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " pod="openstack/ceilometer-0" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.779518 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88c9db22-be72-4ff3-a4bf-361d634922b0" path="/var/lib/kubelet/pods/88c9db22-be72-4ff3-a4bf-361d634922b0/volumes" Dec 03 17:37:33 crc kubenswrapper[4787]: I1203 17:37:33.780192 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="956e5a96-03a5-40ca-a9ce-fbdc8631edb3" path="/var/lib/kubelet/pods/956e5a96-03a5-40ca-a9ce-fbdc8631edb3/volumes" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.105237 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.125630 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.268397 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-scripts\") pod \"e2c6617d-aba8-4b39-94b1-c912a95aa352\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.268469 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-run-httpd\") pod \"e2c6617d-aba8-4b39-94b1-c912a95aa352\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.268734 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-config-data\") pod \"e2c6617d-aba8-4b39-94b1-c912a95aa352\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.268765 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-log-httpd\") pod \"e2c6617d-aba8-4b39-94b1-c912a95aa352\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.268801 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-combined-ca-bundle\") pod \"e2c6617d-aba8-4b39-94b1-c912a95aa352\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.268824 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-sg-core-conf-yaml\") pod \"e2c6617d-aba8-4b39-94b1-c912a95aa352\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.268846 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zrgws\" (UniqueName: \"kubernetes.io/projected/e2c6617d-aba8-4b39-94b1-c912a95aa352-kube-api-access-zrgws\") pod \"e2c6617d-aba8-4b39-94b1-c912a95aa352\" (UID: \"e2c6617d-aba8-4b39-94b1-c912a95aa352\") " Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.271523 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e2c6617d-aba8-4b39-94b1-c912a95aa352" (UID: "e2c6617d-aba8-4b39-94b1-c912a95aa352"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.272666 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2c6617d-aba8-4b39-94b1-c912a95aa352-kube-api-access-zrgws" (OuterVolumeSpecName: "kube-api-access-zrgws") pod "e2c6617d-aba8-4b39-94b1-c912a95aa352" (UID: "e2c6617d-aba8-4b39-94b1-c912a95aa352"). InnerVolumeSpecName "kube-api-access-zrgws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.274705 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e2c6617d-aba8-4b39-94b1-c912a95aa352" (UID: "e2c6617d-aba8-4b39-94b1-c912a95aa352"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.276215 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e2c6617d-aba8-4b39-94b1-c912a95aa352" (UID: "e2c6617d-aba8-4b39-94b1-c912a95aa352"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.276726 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-scripts" (OuterVolumeSpecName: "scripts") pod "e2c6617d-aba8-4b39-94b1-c912a95aa352" (UID: "e2c6617d-aba8-4b39-94b1-c912a95aa352"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.278151 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-config-data" (OuterVolumeSpecName: "config-data") pod "e2c6617d-aba8-4b39-94b1-c912a95aa352" (UID: "e2c6617d-aba8-4b39-94b1-c912a95aa352"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.291644 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2c6617d-aba8-4b39-94b1-c912a95aa352" (UID: "e2c6617d-aba8-4b39-94b1-c912a95aa352"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.371735 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.371775 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.371788 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.371800 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zrgws\" (UniqueName: \"kubernetes.io/projected/e2c6617d-aba8-4b39-94b1-c912a95aa352-kube-api-access-zrgws\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.371811 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.371821 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e2c6617d-aba8-4b39-94b1-c912a95aa352-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.371831 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2c6617d-aba8-4b39-94b1-c912a95aa352-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.754437 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:37:34 crc kubenswrapper[4787]: I1203 17:37:34.813174 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-64cb8b689b-d4pmz"] Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.080212 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.151620 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.151664 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"ff54f756-3848-48ec-a235-d2814ff8d7f8","Type":"ContainerStarted","Data":"a4e540f95c652dc075d471fdf7336d33d079deff1b0b08077f742ad231a6e75d"} Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.194115 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7994c4bbc4-2kvbc"] Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.226673 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.852330974 podStartE2EDuration="37.226650516s" podCreationTimestamp="2025-12-03 17:36:58 +0000 UTC" firstStartedPulling="2025-12-03 17:36:59.859072735 +0000 UTC m=+1436.676543694" lastFinishedPulling="2025-12-03 17:37:34.233392277 +0000 UTC m=+1471.050863236" observedRunningTime="2025-12-03 17:37:35.18655489 +0000 UTC m=+1472.004025849" watchObservedRunningTime="2025-12-03 17:37:35.226650516 +0000 UTC m=+1472.044121475" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.355429 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.367290 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.395708 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.400451 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.404279 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.405108 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.410753 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.431904 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.498936 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.499722 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data-custom\") pod \"8f12c699-048b-4eee-bc7c-e3ad916656ed\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.499858 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data\") pod \"8f12c699-048b-4eee-bc7c-e3ad916656ed\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.499962 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-combined-ca-bundle\") pod \"8f12c699-048b-4eee-bc7c-e3ad916656ed\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.500010 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7z7v\" (UniqueName: \"kubernetes.io/projected/8f12c699-048b-4eee-bc7c-e3ad916656ed-kube-api-access-d7z7v\") pod \"8f12c699-048b-4eee-bc7c-e3ad916656ed\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.500623 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.500707 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-run-httpd\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.500728 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.500773 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-log-httpd\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.500888 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-52mgc\" (UniqueName: \"kubernetes.io/projected/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-kube-api-access-52mgc\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.500918 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-config-data\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.500950 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-scripts\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.509478 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8f12c699-048b-4eee-bc7c-e3ad916656ed" (UID: "8f12c699-048b-4eee-bc7c-e3ad916656ed"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.524984 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f12c699-048b-4eee-bc7c-e3ad916656ed-kube-api-access-d7z7v" (OuterVolumeSpecName: "kube-api-access-d7z7v") pod "8f12c699-048b-4eee-bc7c-e3ad916656ed" (UID: "8f12c699-048b-4eee-bc7c-e3ad916656ed"). InnerVolumeSpecName "kube-api-access-d7z7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.552865 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-8555854d45-f7p99"] Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.553116 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-8555854d45-f7p99" podUID="d7cb720c-c038-4d14-9e53-293022882fcf" containerName="heat-engine" containerID="cri-o://1ceb46625355ff189730f2313117ef55d982bb0f121844b31fa0043239dee766" gracePeriod=60 Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.611640 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f12c699-048b-4eee-bc7c-e3ad916656ed" (UID: "8f12c699-048b-4eee-bc7c-e3ad916656ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.616343 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-combined-ca-bundle\") pod \"8f12c699-048b-4eee-bc7c-e3ad916656ed\" (UID: \"8f12c699-048b-4eee-bc7c-e3ad916656ed\") " Dec 03 17:37:35 crc kubenswrapper[4787]: W1203 17:37:35.616463 4787 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/8f12c699-048b-4eee-bc7c-e3ad916656ed/volumes/kubernetes.io~secret/combined-ca-bundle Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.616480 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8f12c699-048b-4eee-bc7c-e3ad916656ed" (UID: "8f12c699-048b-4eee-bc7c-e3ad916656ed"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.624990 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-log-httpd\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.625313 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-52mgc\" (UniqueName: \"kubernetes.io/projected/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-kube-api-access-52mgc\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.625358 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-config-data\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.625412 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-scripts\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.625464 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.625601 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-run-httpd\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.625617 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.625741 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.625754 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.625786 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7z7v\" (UniqueName: \"kubernetes.io/projected/8f12c699-048b-4eee-bc7c-e3ad916656ed-kube-api-access-d7z7v\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.627721 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-run-httpd\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.632723 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-log-httpd\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.636716 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-config-data\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.643655 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.643671 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.649336 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-scripts\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.661233 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data" (OuterVolumeSpecName: "config-data") pod "8f12c699-048b-4eee-bc7c-e3ad916656ed" (UID: "8f12c699-048b-4eee-bc7c-e3ad916656ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.666076 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-52mgc\" (UniqueName: \"kubernetes.io/projected/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-kube-api-access-52mgc\") pod \"ceilometer-0\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.719555 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.728115 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8f12c699-048b-4eee-bc7c-e3ad916656ed-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.747004 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.780939 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2c6617d-aba8-4b39-94b1-c912a95aa352" path="/var/lib/kubelet/pods/e2c6617d-aba8-4b39-94b1-c912a95aa352/volumes" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.833658 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p27gf\" (UniqueName: \"kubernetes.io/projected/9296f967-7bf3-4e42-a867-cce34092dc11-kube-api-access-p27gf\") pod \"9296f967-7bf3-4e42-a867-cce34092dc11\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.833741 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data\") pod \"9296f967-7bf3-4e42-a867-cce34092dc11\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.833870 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data-custom\") pod \"9296f967-7bf3-4e42-a867-cce34092dc11\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.834001 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-combined-ca-bundle\") pod \"9296f967-7bf3-4e42-a867-cce34092dc11\" (UID: \"9296f967-7bf3-4e42-a867-cce34092dc11\") " Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.840952 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9296f967-7bf3-4e42-a867-cce34092dc11-kube-api-access-p27gf" (OuterVolumeSpecName: "kube-api-access-p27gf") pod "9296f967-7bf3-4e42-a867-cce34092dc11" (UID: "9296f967-7bf3-4e42-a867-cce34092dc11"). InnerVolumeSpecName "kube-api-access-p27gf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.841689 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9296f967-7bf3-4e42-a867-cce34092dc11" (UID: "9296f967-7bf3-4e42-a867-cce34092dc11"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.872212 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9296f967-7bf3-4e42-a867-cce34092dc11" (UID: "9296f967-7bf3-4e42-a867-cce34092dc11"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.932128 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data" (OuterVolumeSpecName: "config-data") pod "9296f967-7bf3-4e42-a867-cce34092dc11" (UID: "9296f967-7bf3-4e42-a867-cce34092dc11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.936523 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p27gf\" (UniqueName: \"kubernetes.io/projected/9296f967-7bf3-4e42-a867-cce34092dc11-kube-api-access-p27gf\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.936554 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.936566 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:35 crc kubenswrapper[4787]: I1203 17:37:35.936577 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9296f967-7bf3-4e42-a867-cce34092dc11-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:36 crc kubenswrapper[4787]: I1203 17:37:36.163122 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" event={"ID":"8f12c699-048b-4eee-bc7c-e3ad916656ed","Type":"ContainerDied","Data":"6b1ad6375492c45a59dc60c5904a65d3dd7e8f5b584633cd82fae68218449a4e"} Dec 03 17:37:36 crc kubenswrapper[4787]: I1203 17:37:36.163179 4787 scope.go:117] "RemoveContainer" containerID="7e594040c2aee27dd4dac76f8492ed0427c1456eebe85855a32e5f68443f301a" Dec 03 17:37:36 crc kubenswrapper[4787]: I1203 17:37:36.163294 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-64cb8b689b-d4pmz" Dec 03 17:37:36 crc kubenswrapper[4787]: I1203 17:37:36.166763 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-7994c4bbc4-2kvbc" event={"ID":"9296f967-7bf3-4e42-a867-cce34092dc11","Type":"ContainerDied","Data":"84bbcb5aaef296b8c23267f71f9f80c12f357655836ee1e36d97193d0e414cc3"} Dec 03 17:37:36 crc kubenswrapper[4787]: I1203 17:37:36.166860 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-7994c4bbc4-2kvbc" Dec 03 17:37:36 crc kubenswrapper[4787]: I1203 17:37:36.197072 4787 scope.go:117] "RemoveContainer" containerID="916e6bd263c38203861c58ce0127628fa60d0530c78a9d5fab6c5aa9394d2f5c" Dec 03 17:37:36 crc kubenswrapper[4787]: I1203 17:37:36.201223 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-64cb8b689b-d4pmz"] Dec 03 17:37:36 crc kubenswrapper[4787]: I1203 17:37:36.216700 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-64cb8b689b-d4pmz"] Dec 03 17:37:36 crc kubenswrapper[4787]: I1203 17:37:36.226931 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-7994c4bbc4-2kvbc"] Dec 03 17:37:36 crc kubenswrapper[4787]: I1203 17:37:36.240749 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-7994c4bbc4-2kvbc"] Dec 03 17:37:36 crc kubenswrapper[4787]: I1203 17:37:36.274542 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:36 crc kubenswrapper[4787]: E1203 17:37:36.569514 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ceb46625355ff189730f2313117ef55d982bb0f121844b31fa0043239dee766" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:37:36 crc kubenswrapper[4787]: E1203 17:37:36.571620 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ceb46625355ff189730f2313117ef55d982bb0f121844b31fa0043239dee766" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:37:36 crc kubenswrapper[4787]: E1203 17:37:36.572688 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ceb46625355ff189730f2313117ef55d982bb0f121844b31fa0043239dee766" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:37:36 crc kubenswrapper[4787]: E1203 17:37:36.572727 4787 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-8555854d45-f7p99" podUID="d7cb720c-c038-4d14-9e53-293022882fcf" containerName="heat-engine" Dec 03 17:37:37 crc kubenswrapper[4787]: I1203 17:37:37.180957 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4","Type":"ContainerStarted","Data":"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a"} Dec 03 17:37:37 crc kubenswrapper[4787]: I1203 17:37:37.181276 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4","Type":"ContainerStarted","Data":"eaa28622f476f134a1b98d24249447391a4ba09f13cee6f25e8953df9b6322e6"} Dec 03 17:37:37 crc kubenswrapper[4787]: I1203 17:37:37.780525 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f12c699-048b-4eee-bc7c-e3ad916656ed" path="/var/lib/kubelet/pods/8f12c699-048b-4eee-bc7c-e3ad916656ed/volumes" Dec 03 17:37:37 crc kubenswrapper[4787]: I1203 17:37:37.781818 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9296f967-7bf3-4e42-a867-cce34092dc11" path="/var/lib/kubelet/pods/9296f967-7bf3-4e42-a867-cce34092dc11/volumes" Dec 03 17:37:40 crc kubenswrapper[4787]: I1203 17:37:40.356884 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 03 17:37:44 crc kubenswrapper[4787]: I1203 17:37:44.276708 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4","Type":"ContainerStarted","Data":"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8"} Dec 03 17:37:44 crc kubenswrapper[4787]: I1203 17:37:44.279856 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l2pkv" event={"ID":"131d89aa-86f5-4ddb-a388-f2a938319ce2","Type":"ContainerStarted","Data":"4e810cda512274f9a0e0575876cb1477e0a06af4d0c0f1d365cda6ba6262f8fb"} Dec 03 17:37:44 crc kubenswrapper[4787]: I1203 17:37:44.308784 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-l2pkv" podStartSLOduration=2.109198425 podStartE2EDuration="14.308764194s" podCreationTimestamp="2025-12-03 17:37:30 +0000 UTC" firstStartedPulling="2025-12-03 17:37:31.583693959 +0000 UTC m=+1468.401164918" lastFinishedPulling="2025-12-03 17:37:43.783259728 +0000 UTC m=+1480.600730687" observedRunningTime="2025-12-03 17:37:44.296028399 +0000 UTC m=+1481.113499378" watchObservedRunningTime="2025-12-03 17:37:44.308764194 +0000 UTC m=+1481.126235153" Dec 03 17:37:46 crc kubenswrapper[4787]: I1203 17:37:46.308522 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4","Type":"ContainerStarted","Data":"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b"} Dec 03 17:37:46 crc kubenswrapper[4787]: E1203 17:37:46.568665 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ceb46625355ff189730f2313117ef55d982bb0f121844b31fa0043239dee766" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:37:46 crc kubenswrapper[4787]: E1203 17:37:46.572734 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ceb46625355ff189730f2313117ef55d982bb0f121844b31fa0043239dee766" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:37:46 crc kubenswrapper[4787]: E1203 17:37:46.574776 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="1ceb46625355ff189730f2313117ef55d982bb0f121844b31fa0043239dee766" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:37:46 crc kubenswrapper[4787]: E1203 17:37:46.574848 4787 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-8555854d45-f7p99" podUID="d7cb720c-c038-4d14-9e53-293022882fcf" containerName="heat-engine" Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.322493 4787 generic.go:334] "Generic (PLEG): container finished" podID="d7cb720c-c038-4d14-9e53-293022882fcf" containerID="1ceb46625355ff189730f2313117ef55d982bb0f121844b31fa0043239dee766" exitCode=0 Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.322586 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-8555854d45-f7p99" event={"ID":"d7cb720c-c038-4d14-9e53-293022882fcf","Type":"ContainerDied","Data":"1ceb46625355ff189730f2313117ef55d982bb0f121844b31fa0043239dee766"} Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.394735 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.506719 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data\") pod \"d7cb720c-c038-4d14-9e53-293022882fcf\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.506900 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-combined-ca-bundle\") pod \"d7cb720c-c038-4d14-9e53-293022882fcf\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.506984 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data-custom\") pod \"d7cb720c-c038-4d14-9e53-293022882fcf\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.507066 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jqkpr\" (UniqueName: \"kubernetes.io/projected/d7cb720c-c038-4d14-9e53-293022882fcf-kube-api-access-jqkpr\") pod \"d7cb720c-c038-4d14-9e53-293022882fcf\" (UID: \"d7cb720c-c038-4d14-9e53-293022882fcf\") " Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.515911 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d7cb720c-c038-4d14-9e53-293022882fcf-kube-api-access-jqkpr" (OuterVolumeSpecName: "kube-api-access-jqkpr") pod "d7cb720c-c038-4d14-9e53-293022882fcf" (UID: "d7cb720c-c038-4d14-9e53-293022882fcf"). InnerVolumeSpecName "kube-api-access-jqkpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.518314 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "d7cb720c-c038-4d14-9e53-293022882fcf" (UID: "d7cb720c-c038-4d14-9e53-293022882fcf"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.548091 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d7cb720c-c038-4d14-9e53-293022882fcf" (UID: "d7cb720c-c038-4d14-9e53-293022882fcf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.583709 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data" (OuterVolumeSpecName: "config-data") pod "d7cb720c-c038-4d14-9e53-293022882fcf" (UID: "d7cb720c-c038-4d14-9e53-293022882fcf"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.613584 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.613629 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.613644 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/d7cb720c-c038-4d14-9e53-293022882fcf-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:47 crc kubenswrapper[4787]: I1203 17:37:47.613656 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jqkpr\" (UniqueName: \"kubernetes.io/projected/d7cb720c-c038-4d14-9e53-293022882fcf-kube-api-access-jqkpr\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.306966 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.338909 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4","Type":"ContainerStarted","Data":"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7"} Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.339959 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.342902 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-8555854d45-f7p99" event={"ID":"d7cb720c-c038-4d14-9e53-293022882fcf","Type":"ContainerDied","Data":"3283fe2bd2aa7f36942c05618ef501b2d91323d5979fa56f988dc484f1459932"} Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.342956 4787 scope.go:117] "RemoveContainer" containerID="1ceb46625355ff189730f2313117ef55d982bb0f121844b31fa0043239dee766" Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.343112 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-8555854d45-f7p99" Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.371389 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.617147858 podStartE2EDuration="13.371369191s" podCreationTimestamp="2025-12-03 17:37:35 +0000 UTC" firstStartedPulling="2025-12-03 17:37:36.283875826 +0000 UTC m=+1473.101346785" lastFinishedPulling="2025-12-03 17:37:47.038097159 +0000 UTC m=+1483.855568118" observedRunningTime="2025-12-03 17:37:48.365966924 +0000 UTC m=+1485.183437913" watchObservedRunningTime="2025-12-03 17:37:48.371369191 +0000 UTC m=+1485.188840170" Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.397587 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-8555854d45-f7p99"] Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.413577 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-8555854d45-f7p99"] Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.990281 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.990330 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.990370 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.991397 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8530351cb9d4d0e6acbfd6734059c845290cade8ecc29a25ad6ed13ae5d5db93"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:37:48 crc kubenswrapper[4787]: I1203 17:37:48.991456 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://8530351cb9d4d0e6acbfd6734059c845290cade8ecc29a25ad6ed13ae5d5db93" gracePeriod=600 Dec 03 17:37:49 crc kubenswrapper[4787]: I1203 17:37:49.355961 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="8530351cb9d4d0e6acbfd6734059c845290cade8ecc29a25ad6ed13ae5d5db93" exitCode=0 Dec 03 17:37:49 crc kubenswrapper[4787]: I1203 17:37:49.356064 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"8530351cb9d4d0e6acbfd6734059c845290cade8ecc29a25ad6ed13ae5d5db93"} Dec 03 17:37:49 crc kubenswrapper[4787]: I1203 17:37:49.356395 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75"} Dec 03 17:37:49 crc kubenswrapper[4787]: I1203 17:37:49.356417 4787 scope.go:117] "RemoveContainer" containerID="93e6b55e4dd0c75878fc7a5cbe5a33dad37bb97aa5cbfc657a6c1fd769073445" Dec 03 17:37:49 crc kubenswrapper[4787]: I1203 17:37:49.357957 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="ceilometer-central-agent" containerID="cri-o://4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a" gracePeriod=30 Dec 03 17:37:49 crc kubenswrapper[4787]: I1203 17:37:49.358109 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="proxy-httpd" containerID="cri-o://6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7" gracePeriod=30 Dec 03 17:37:49 crc kubenswrapper[4787]: I1203 17:37:49.358148 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="sg-core" containerID="cri-o://9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b" gracePeriod=30 Dec 03 17:37:49 crc kubenswrapper[4787]: I1203 17:37:49.358178 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="ceilometer-notification-agent" containerID="cri-o://2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8" gracePeriod=30 Dec 03 17:37:49 crc kubenswrapper[4787]: I1203 17:37:49.794384 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d7cb720c-c038-4d14-9e53-293022882fcf" path="/var/lib/kubelet/pods/d7cb720c-c038-4d14-9e53-293022882fcf/volumes" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.359373 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.370485 4787 generic.go:334] "Generic (PLEG): container finished" podID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerID="6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7" exitCode=0 Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.370520 4787 generic.go:334] "Generic (PLEG): container finished" podID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerID="9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b" exitCode=2 Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.370530 4787 generic.go:334] "Generic (PLEG): container finished" podID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerID="2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8" exitCode=0 Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.370538 4787 generic.go:334] "Generic (PLEG): container finished" podID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerID="4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a" exitCode=0 Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.370599 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4","Type":"ContainerDied","Data":"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7"} Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.370635 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4","Type":"ContainerDied","Data":"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b"} Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.370650 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4","Type":"ContainerDied","Data":"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8"} Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.370662 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4","Type":"ContainerDied","Data":"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a"} Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.370672 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4","Type":"ContainerDied","Data":"eaa28622f476f134a1b98d24249447391a4ba09f13cee6f25e8953df9b6322e6"} Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.370690 4787 scope.go:117] "RemoveContainer" containerID="6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.370818 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.424260 4787 scope.go:117] "RemoveContainer" containerID="9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.461007 4787 scope.go:117] "RemoveContainer" containerID="2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.475145 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-52mgc\" (UniqueName: \"kubernetes.io/projected/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-kube-api-access-52mgc\") pod \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.476073 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-run-httpd\") pod \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.476250 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-sg-core-conf-yaml\") pod \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.476309 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-scripts\") pod \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.476337 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-combined-ca-bundle\") pod \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.476111 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" (UID: "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.476491 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-config-data\") pod \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.476922 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-log-httpd\") pod \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\" (UID: \"59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4\") " Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.477348 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" (UID: "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.477884 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.477956 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.484192 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-kube-api-access-52mgc" (OuterVolumeSpecName: "kube-api-access-52mgc") pod "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" (UID: "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4"). InnerVolumeSpecName "kube-api-access-52mgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.484227 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-scripts" (OuterVolumeSpecName: "scripts") pod "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" (UID: "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.491761 4787 scope.go:117] "RemoveContainer" containerID="4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.510709 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" (UID: "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.579607 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-52mgc\" (UniqueName: \"kubernetes.io/projected/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-kube-api-access-52mgc\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.579640 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.579650 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.580844 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" (UID: "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.602394 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-config-data" (OuterVolumeSpecName: "config-data") pod "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" (UID: "59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.667608 4787 scope.go:117] "RemoveContainer" containerID="6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.668645 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7\": container with ID starting with 6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7 not found: ID does not exist" containerID="6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.668696 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7"} err="failed to get container status \"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7\": rpc error: code = NotFound desc = could not find container \"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7\": container with ID starting with 6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7 not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.668729 4787 scope.go:117] "RemoveContainer" containerID="9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.669147 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b\": container with ID starting with 9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b not found: ID does not exist" containerID="9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.669216 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b"} err="failed to get container status \"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b\": rpc error: code = NotFound desc = could not find container \"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b\": container with ID starting with 9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.669250 4787 scope.go:117] "RemoveContainer" containerID="2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.669568 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8\": container with ID starting with 2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8 not found: ID does not exist" containerID="2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.669597 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8"} err="failed to get container status \"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8\": rpc error: code = NotFound desc = could not find container \"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8\": container with ID starting with 2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8 not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.669617 4787 scope.go:117] "RemoveContainer" containerID="4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.670442 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a\": container with ID starting with 4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a not found: ID does not exist" containerID="4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.670484 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a"} err="failed to get container status \"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a\": rpc error: code = NotFound desc = could not find container \"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a\": container with ID starting with 4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.670505 4787 scope.go:117] "RemoveContainer" containerID="6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.678085 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7"} err="failed to get container status \"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7\": rpc error: code = NotFound desc = could not find container \"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7\": container with ID starting with 6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7 not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.678134 4787 scope.go:117] "RemoveContainer" containerID="9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.678627 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b"} err="failed to get container status \"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b\": rpc error: code = NotFound desc = could not find container \"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b\": container with ID starting with 9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.678742 4787 scope.go:117] "RemoveContainer" containerID="2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.679062 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8"} err="failed to get container status \"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8\": rpc error: code = NotFound desc = could not find container \"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8\": container with ID starting with 2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8 not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.679081 4787 scope.go:117] "RemoveContainer" containerID="4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.679283 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a"} err="failed to get container status \"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a\": rpc error: code = NotFound desc = could not find container \"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a\": container with ID starting with 4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.679307 4787 scope.go:117] "RemoveContainer" containerID="6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.679534 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7"} err="failed to get container status \"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7\": rpc error: code = NotFound desc = could not find container \"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7\": container with ID starting with 6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7 not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.679554 4787 scope.go:117] "RemoveContainer" containerID="9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.679825 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b"} err="failed to get container status \"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b\": rpc error: code = NotFound desc = could not find container \"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b\": container with ID starting with 9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.679854 4787 scope.go:117] "RemoveContainer" containerID="2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.681079 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8"} err="failed to get container status \"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8\": rpc error: code = NotFound desc = could not find container \"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8\": container with ID starting with 2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8 not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.681133 4787 scope.go:117] "RemoveContainer" containerID="4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.681144 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.681161 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.681591 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a"} err="failed to get container status \"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a\": rpc error: code = NotFound desc = could not find container \"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a\": container with ID starting with 4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.681619 4787 scope.go:117] "RemoveContainer" containerID="6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.681942 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7"} err="failed to get container status \"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7\": rpc error: code = NotFound desc = could not find container \"6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7\": container with ID starting with 6e2a791a8e29479c8ff34162c29b6cf4d9255bcac408ad41e133e5009badc7a7 not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.681971 4787 scope.go:117] "RemoveContainer" containerID="9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.682165 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b"} err="failed to get container status \"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b\": rpc error: code = NotFound desc = could not find container \"9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b\": container with ID starting with 9e203b3064944a2af38d10244d121d779278bebb77a2d2e5d8de4ce5eaaca09b not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.682189 4787 scope.go:117] "RemoveContainer" containerID="2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.682387 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8"} err="failed to get container status \"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8\": rpc error: code = NotFound desc = could not find container \"2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8\": container with ID starting with 2adc74f3fef1ad1562675b8afa84dcc62a5c360df34b126367429d4d5af11dc8 not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.682411 4787 scope.go:117] "RemoveContainer" containerID="4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.682613 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a"} err="failed to get container status \"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a\": rpc error: code = NotFound desc = could not find container \"4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a\": container with ID starting with 4581d2a463ea19c662ab24c780309e29984c71bd69927d97ef82c20c47ab118a not found: ID does not exist" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.711603 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.720246 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.746350 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.747078 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9296f967-7bf3-4e42-a867-cce34092dc11" containerName="heat-api" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747097 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9296f967-7bf3-4e42-a867-cce34092dc11" containerName="heat-api" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.747116 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9296f967-7bf3-4e42-a867-cce34092dc11" containerName="heat-api" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747122 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9296f967-7bf3-4e42-a867-cce34092dc11" containerName="heat-api" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.747134 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="ceilometer-notification-agent" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747140 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="ceilometer-notification-agent" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.747148 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="ceilometer-central-agent" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747153 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="ceilometer-central-agent" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.747170 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d7cb720c-c038-4d14-9e53-293022882fcf" containerName="heat-engine" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747175 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d7cb720c-c038-4d14-9e53-293022882fcf" containerName="heat-engine" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.747191 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f12c699-048b-4eee-bc7c-e3ad916656ed" containerName="heat-cfnapi" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747196 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f12c699-048b-4eee-bc7c-e3ad916656ed" containerName="heat-cfnapi" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.747207 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f12c699-048b-4eee-bc7c-e3ad916656ed" containerName="heat-cfnapi" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747212 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f12c699-048b-4eee-bc7c-e3ad916656ed" containerName="heat-cfnapi" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.747231 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="proxy-httpd" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747239 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="proxy-httpd" Dec 03 17:37:50 crc kubenswrapper[4787]: E1203 17:37:50.747246 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="sg-core" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747252 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="sg-core" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747427 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f12c699-048b-4eee-bc7c-e3ad916656ed" containerName="heat-cfnapi" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747449 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9296f967-7bf3-4e42-a867-cce34092dc11" containerName="heat-api" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747463 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d7cb720c-c038-4d14-9e53-293022882fcf" containerName="heat-engine" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747471 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="sg-core" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747479 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="ceilometer-notification-agent" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747488 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="proxy-httpd" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747498 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f12c699-048b-4eee-bc7c-e3ad916656ed" containerName="heat-cfnapi" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747514 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" containerName="ceilometer-central-agent" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.747896 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9296f967-7bf3-4e42-a867-cce34092dc11" containerName="heat-api" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.749288 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.751781 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.751948 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.769689 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.899898 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-config-data\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.899959 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-run-httpd\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.899993 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf986\" (UniqueName: \"kubernetes.io/projected/b73232f2-6a8f-4008-b85a-5abde10afac8-kube-api-access-wf986\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.900034 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-log-httpd\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.900066 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-scripts\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.900125 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:50 crc kubenswrapper[4787]: I1203 17:37:50.900203 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.001705 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-config-data\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.002120 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-run-httpd\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.002153 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf986\" (UniqueName: \"kubernetes.io/projected/b73232f2-6a8f-4008-b85a-5abde10afac8-kube-api-access-wf986\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.002183 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-log-httpd\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.002216 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-scripts\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.002274 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.002373 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.002659 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-run-httpd\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.002852 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-log-httpd\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.007958 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-scripts\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.008353 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.008441 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.010533 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-config-data\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.023700 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf986\" (UniqueName: \"kubernetes.io/projected/b73232f2-6a8f-4008-b85a-5abde10afac8-kube-api-access-wf986\") pod \"ceilometer-0\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.071364 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.546794 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:37:51 crc kubenswrapper[4787]: W1203 17:37:51.553635 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb73232f2_6a8f_4008_b85a_5abde10afac8.slice/crio-d01fc51856721f67159803f7287792195d5bbda1df369a666aaa4c1dfac98cea WatchSource:0}: Error finding container d01fc51856721f67159803f7287792195d5bbda1df369a666aaa4c1dfac98cea: Status 404 returned error can't find the container with id d01fc51856721f67159803f7287792195d5bbda1df369a666aaa4c1dfac98cea Dec 03 17:37:51 crc kubenswrapper[4787]: I1203 17:37:51.782855 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4" path="/var/lib/kubelet/pods/59bbc9c8-3bd5-4e50-a13a-e3b91a06e9c4/volumes" Dec 03 17:37:52 crc kubenswrapper[4787]: I1203 17:37:52.404416 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b73232f2-6a8f-4008-b85a-5abde10afac8","Type":"ContainerStarted","Data":"d01fc51856721f67159803f7287792195d5bbda1df369a666aaa4c1dfac98cea"} Dec 03 17:37:53 crc kubenswrapper[4787]: I1203 17:37:53.417610 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b73232f2-6a8f-4008-b85a-5abde10afac8","Type":"ContainerStarted","Data":"93dd278d59121f8b452915cd6cddb7ad9722212416513b86140fcbd6c969f270"} Dec 03 17:37:54 crc kubenswrapper[4787]: I1203 17:37:54.463740 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b73232f2-6a8f-4008-b85a-5abde10afac8","Type":"ContainerStarted","Data":"866ac7ebfe6c706fe719e1b67dbbe6c90fcc36d8f86913e74ee9d1fe8f2b5f75"} Dec 03 17:37:55 crc kubenswrapper[4787]: I1203 17:37:55.480593 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b73232f2-6a8f-4008-b85a-5abde10afac8","Type":"ContainerStarted","Data":"7d95e709d11a7d7a5fb923a98562238a3910bfc6ece04a7d257edc44eb8008ef"} Dec 03 17:37:56 crc kubenswrapper[4787]: I1203 17:37:56.493664 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b73232f2-6a8f-4008-b85a-5abde10afac8","Type":"ContainerStarted","Data":"7437a6ffa988eb880e4d2bd09c867a8252afe9a4495921f4c650e8d470c59b46"} Dec 03 17:37:56 crc kubenswrapper[4787]: I1203 17:37:56.493972 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:37:56 crc kubenswrapper[4787]: I1203 17:37:56.522779 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.236826401 podStartE2EDuration="6.522757494s" podCreationTimestamp="2025-12-03 17:37:50 +0000 UTC" firstStartedPulling="2025-12-03 17:37:51.556776841 +0000 UTC m=+1488.374247800" lastFinishedPulling="2025-12-03 17:37:55.842707934 +0000 UTC m=+1492.660178893" observedRunningTime="2025-12-03 17:37:56.515376394 +0000 UTC m=+1493.332847393" watchObservedRunningTime="2025-12-03 17:37:56.522757494 +0000 UTC m=+1493.340228453" Dec 03 17:37:59 crc kubenswrapper[4787]: I1203 17:37:59.533061 4787 generic.go:334] "Generic (PLEG): container finished" podID="131d89aa-86f5-4ddb-a388-f2a938319ce2" containerID="4e810cda512274f9a0e0575876cb1477e0a06af4d0c0f1d365cda6ba6262f8fb" exitCode=0 Dec 03 17:37:59 crc kubenswrapper[4787]: I1203 17:37:59.533166 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l2pkv" event={"ID":"131d89aa-86f5-4ddb-a388-f2a938319ce2","Type":"ContainerDied","Data":"4e810cda512274f9a0e0575876cb1477e0a06af4d0c0f1d365cda6ba6262f8fb"} Dec 03 17:38:00 crc kubenswrapper[4787]: I1203 17:38:00.980925 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.137893 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nfqm\" (UniqueName: \"kubernetes.io/projected/131d89aa-86f5-4ddb-a388-f2a938319ce2-kube-api-access-7nfqm\") pod \"131d89aa-86f5-4ddb-a388-f2a938319ce2\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.137955 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-combined-ca-bundle\") pod \"131d89aa-86f5-4ddb-a388-f2a938319ce2\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.138104 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-config-data\") pod \"131d89aa-86f5-4ddb-a388-f2a938319ce2\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.138234 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-scripts\") pod \"131d89aa-86f5-4ddb-a388-f2a938319ce2\" (UID: \"131d89aa-86f5-4ddb-a388-f2a938319ce2\") " Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.145056 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-scripts" (OuterVolumeSpecName: "scripts") pod "131d89aa-86f5-4ddb-a388-f2a938319ce2" (UID: "131d89aa-86f5-4ddb-a388-f2a938319ce2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.147522 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/131d89aa-86f5-4ddb-a388-f2a938319ce2-kube-api-access-7nfqm" (OuterVolumeSpecName: "kube-api-access-7nfqm") pod "131d89aa-86f5-4ddb-a388-f2a938319ce2" (UID: "131d89aa-86f5-4ddb-a388-f2a938319ce2"). InnerVolumeSpecName "kube-api-access-7nfqm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.174250 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-config-data" (OuterVolumeSpecName: "config-data") pod "131d89aa-86f5-4ddb-a388-f2a938319ce2" (UID: "131d89aa-86f5-4ddb-a388-f2a938319ce2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.196764 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "131d89aa-86f5-4ddb-a388-f2a938319ce2" (UID: "131d89aa-86f5-4ddb-a388-f2a938319ce2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.240568 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.240770 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nfqm\" (UniqueName: \"kubernetes.io/projected/131d89aa-86f5-4ddb-a388-f2a938319ce2-kube-api-access-7nfqm\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.240868 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.240934 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/131d89aa-86f5-4ddb-a388-f2a938319ce2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.557519 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-l2pkv" event={"ID":"131d89aa-86f5-4ddb-a388-f2a938319ce2","Type":"ContainerDied","Data":"8c5610c461e1449247e5339eb15bc2ce56923144754d4a53c718ab3759ceeb62"} Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.557568 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8c5610c461e1449247e5339eb15bc2ce56923144754d4a53c718ab3759ceeb62" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.557571 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-l2pkv" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.667840 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:38:01 crc kubenswrapper[4787]: E1203 17:38:01.668531 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="131d89aa-86f5-4ddb-a388-f2a938319ce2" containerName="nova-cell0-conductor-db-sync" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.668615 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="131d89aa-86f5-4ddb-a388-f2a938319ce2" containerName="nova-cell0-conductor-db-sync" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.668911 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="131d89aa-86f5-4ddb-a388-f2a938319ce2" containerName="nova-cell0-conductor-db-sync" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.669718 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.671705 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-669qv" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.672911 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.706201 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.751869 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.752049 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.752120 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pw7xw\" (UniqueName: \"kubernetes.io/projected/ac1922ce-f523-4e05-b9c1-676df04be85d-kube-api-access-pw7xw\") pod \"nova-cell0-conductor-0\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.853843 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.854071 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.854180 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pw7xw\" (UniqueName: \"kubernetes.io/projected/ac1922ce-f523-4e05-b9c1-676df04be85d-kube-api-access-pw7xw\") pod \"nova-cell0-conductor-0\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.863651 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.864951 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.878498 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pw7xw\" (UniqueName: \"kubernetes.io/projected/ac1922ce-f523-4e05-b9c1-676df04be85d-kube-api-access-pw7xw\") pod \"nova-cell0-conductor-0\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.882834 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.883210 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="ceilometer-central-agent" containerID="cri-o://93dd278d59121f8b452915cd6cddb7ad9722212416513b86140fcbd6c969f270" gracePeriod=30 Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.883267 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="sg-core" containerID="cri-o://7d95e709d11a7d7a5fb923a98562238a3910bfc6ece04a7d257edc44eb8008ef" gracePeriod=30 Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.883362 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="ceilometer-notification-agent" containerID="cri-o://866ac7ebfe6c706fe719e1b67dbbe6c90fcc36d8f86913e74ee9d1fe8f2b5f75" gracePeriod=30 Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.883392 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="proxy-httpd" containerID="cri-o://7437a6ffa988eb880e4d2bd09c867a8252afe9a4495921f4c650e8d470c59b46" gracePeriod=30 Dec 03 17:38:01 crc kubenswrapper[4787]: I1203 17:38:01.989048 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:02 crc kubenswrapper[4787]: I1203 17:38:02.479282 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:38:02 crc kubenswrapper[4787]: I1203 17:38:02.569251 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ac1922ce-f523-4e05-b9c1-676df04be85d","Type":"ContainerStarted","Data":"93e2a4cbd36696d9e768224e4e1788161d246595e77ae2e97f6ae79a73d6401b"} Dec 03 17:38:02 crc kubenswrapper[4787]: I1203 17:38:02.573996 4787 generic.go:334] "Generic (PLEG): container finished" podID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerID="7437a6ffa988eb880e4d2bd09c867a8252afe9a4495921f4c650e8d470c59b46" exitCode=0 Dec 03 17:38:02 crc kubenswrapper[4787]: I1203 17:38:02.574056 4787 generic.go:334] "Generic (PLEG): container finished" podID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerID="7d95e709d11a7d7a5fb923a98562238a3910bfc6ece04a7d257edc44eb8008ef" exitCode=2 Dec 03 17:38:02 crc kubenswrapper[4787]: I1203 17:38:02.574067 4787 generic.go:334] "Generic (PLEG): container finished" podID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerID="866ac7ebfe6c706fe719e1b67dbbe6c90fcc36d8f86913e74ee9d1fe8f2b5f75" exitCode=0 Dec 03 17:38:02 crc kubenswrapper[4787]: I1203 17:38:02.574056 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b73232f2-6a8f-4008-b85a-5abde10afac8","Type":"ContainerDied","Data":"7437a6ffa988eb880e4d2bd09c867a8252afe9a4495921f4c650e8d470c59b46"} Dec 03 17:38:02 crc kubenswrapper[4787]: I1203 17:38:02.574111 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b73232f2-6a8f-4008-b85a-5abde10afac8","Type":"ContainerDied","Data":"7d95e709d11a7d7a5fb923a98562238a3910bfc6ece04a7d257edc44eb8008ef"} Dec 03 17:38:02 crc kubenswrapper[4787]: I1203 17:38:02.574133 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b73232f2-6a8f-4008-b85a-5abde10afac8","Type":"ContainerDied","Data":"866ac7ebfe6c706fe719e1b67dbbe6c90fcc36d8f86913e74ee9d1fe8f2b5f75"} Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.590973 4787 generic.go:334] "Generic (PLEG): container finished" podID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerID="93dd278d59121f8b452915cd6cddb7ad9722212416513b86140fcbd6c969f270" exitCode=0 Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.591185 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b73232f2-6a8f-4008-b85a-5abde10afac8","Type":"ContainerDied","Data":"93dd278d59121f8b452915cd6cddb7ad9722212416513b86140fcbd6c969f270"} Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.592964 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b73232f2-6a8f-4008-b85a-5abde10afac8","Type":"ContainerDied","Data":"d01fc51856721f67159803f7287792195d5bbda1df369a666aaa4c1dfac98cea"} Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.593162 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d01fc51856721f67159803f7287792195d5bbda1df369a666aaa4c1dfac98cea" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.595353 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ac1922ce-f523-4e05-b9c1-676df04be85d","Type":"ContainerStarted","Data":"c74d5f2accf20704716e0459f555a75c24a58804dbcda99587a7d7a1f598b8d4"} Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.596938 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.621993 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.621973403 podStartE2EDuration="2.621973403s" podCreationTimestamp="2025-12-03 17:38:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:03.617076131 +0000 UTC m=+1500.434547100" watchObservedRunningTime="2025-12-03 17:38:03.621973403 +0000 UTC m=+1500.439444362" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.662702 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.792042 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-sg-core-conf-yaml\") pod \"b73232f2-6a8f-4008-b85a-5abde10afac8\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.792376 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-log-httpd\") pod \"b73232f2-6a8f-4008-b85a-5abde10afac8\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.792441 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-combined-ca-bundle\") pod \"b73232f2-6a8f-4008-b85a-5abde10afac8\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.792460 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-run-httpd\") pod \"b73232f2-6a8f-4008-b85a-5abde10afac8\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.792539 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-scripts\") pod \"b73232f2-6a8f-4008-b85a-5abde10afac8\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.792653 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-config-data\") pod \"b73232f2-6a8f-4008-b85a-5abde10afac8\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.792700 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wf986\" (UniqueName: \"kubernetes.io/projected/b73232f2-6a8f-4008-b85a-5abde10afac8-kube-api-access-wf986\") pod \"b73232f2-6a8f-4008-b85a-5abde10afac8\" (UID: \"b73232f2-6a8f-4008-b85a-5abde10afac8\") " Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.793533 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b73232f2-6a8f-4008-b85a-5abde10afac8" (UID: "b73232f2-6a8f-4008-b85a-5abde10afac8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.793928 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b73232f2-6a8f-4008-b85a-5abde10afac8" (UID: "b73232f2-6a8f-4008-b85a-5abde10afac8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.798423 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-scripts" (OuterVolumeSpecName: "scripts") pod "b73232f2-6a8f-4008-b85a-5abde10afac8" (UID: "b73232f2-6a8f-4008-b85a-5abde10afac8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.808445 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b73232f2-6a8f-4008-b85a-5abde10afac8-kube-api-access-wf986" (OuterVolumeSpecName: "kube-api-access-wf986") pod "b73232f2-6a8f-4008-b85a-5abde10afac8" (UID: "b73232f2-6a8f-4008-b85a-5abde10afac8"). InnerVolumeSpecName "kube-api-access-wf986". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.845032 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b73232f2-6a8f-4008-b85a-5abde10afac8" (UID: "b73232f2-6a8f-4008-b85a-5abde10afac8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.895676 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.895709 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wf986\" (UniqueName: \"kubernetes.io/projected/b73232f2-6a8f-4008-b85a-5abde10afac8-kube-api-access-wf986\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.895718 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.895727 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.895736 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b73232f2-6a8f-4008-b85a-5abde10afac8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.909516 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b73232f2-6a8f-4008-b85a-5abde10afac8" (UID: "b73232f2-6a8f-4008-b85a-5abde10afac8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.941085 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-config-data" (OuterVolumeSpecName: "config-data") pod "b73232f2-6a8f-4008-b85a-5abde10afac8" (UID: "b73232f2-6a8f-4008-b85a-5abde10afac8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.997304 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:03 crc kubenswrapper[4787]: I1203 17:38:03.997337 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b73232f2-6a8f-4008-b85a-5abde10afac8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.603618 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.656593 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.688409 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.709352 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:04 crc kubenswrapper[4787]: E1203 17:38:04.710599 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="ceilometer-central-agent" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.710624 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="ceilometer-central-agent" Dec 03 17:38:04 crc kubenswrapper[4787]: E1203 17:38:04.710636 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="sg-core" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.710643 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="sg-core" Dec 03 17:38:04 crc kubenswrapper[4787]: E1203 17:38:04.710699 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="proxy-httpd" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.710706 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="proxy-httpd" Dec 03 17:38:04 crc kubenswrapper[4787]: E1203 17:38:04.710736 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="ceilometer-notification-agent" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.710766 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="ceilometer-notification-agent" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.711426 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="sg-core" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.711454 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="proxy-httpd" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.711477 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="ceilometer-notification-agent" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.711487 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" containerName="ceilometer-central-agent" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.716480 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.719571 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.725052 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.755517 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.821055 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-log-httpd\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.821116 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.821146 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.821230 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2djxs\" (UniqueName: \"kubernetes.io/projected/85ab2628-3d14-4c54-8314-1c395e11a71b-kube-api-access-2djxs\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.821276 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-scripts\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.821306 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-run-httpd\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.821328 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-config-data\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.927414 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-log-httpd\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.927478 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.927509 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.927575 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2djxs\" (UniqueName: \"kubernetes.io/projected/85ab2628-3d14-4c54-8314-1c395e11a71b-kube-api-access-2djxs\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.927618 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-scripts\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.927641 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-run-httpd\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.927657 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-config-data\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.930341 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-log-httpd\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.932265 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-run-httpd\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.936284 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.937333 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-scripts\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.943772 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-config-data\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.944092 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:04 crc kubenswrapper[4787]: I1203 17:38:04.949871 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2djxs\" (UniqueName: \"kubernetes.io/projected/85ab2628-3d14-4c54-8314-1c395e11a71b-kube-api-access-2djxs\") pod \"ceilometer-0\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " pod="openstack/ceilometer-0" Dec 03 17:38:05 crc kubenswrapper[4787]: I1203 17:38:05.055209 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:05 crc kubenswrapper[4787]: I1203 17:38:05.535850 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:05 crc kubenswrapper[4787]: W1203 17:38:05.563238 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85ab2628_3d14_4c54_8314_1c395e11a71b.slice/crio-73da631ec08fe842a9d99b12ac0262aec82089975a6faf43b40ae7ac69bdb046 WatchSource:0}: Error finding container 73da631ec08fe842a9d99b12ac0262aec82089975a6faf43b40ae7ac69bdb046: Status 404 returned error can't find the container with id 73da631ec08fe842a9d99b12ac0262aec82089975a6faf43b40ae7ac69bdb046 Dec 03 17:38:05 crc kubenswrapper[4787]: I1203 17:38:05.626286 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ab2628-3d14-4c54-8314-1c395e11a71b","Type":"ContainerStarted","Data":"73da631ec08fe842a9d99b12ac0262aec82089975a6faf43b40ae7ac69bdb046"} Dec 03 17:38:05 crc kubenswrapper[4787]: I1203 17:38:05.779789 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b73232f2-6a8f-4008-b85a-5abde10afac8" path="/var/lib/kubelet/pods/b73232f2-6a8f-4008-b85a-5abde10afac8/volumes" Dec 03 17:38:06 crc kubenswrapper[4787]: I1203 17:38:06.639762 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ab2628-3d14-4c54-8314-1c395e11a71b","Type":"ContainerStarted","Data":"110cb8edea0aaf7f5dd3511f44ed88af2276c4b72f984eaadec63eec6188a897"} Dec 03 17:38:07 crc kubenswrapper[4787]: I1203 17:38:07.656726 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ab2628-3d14-4c54-8314-1c395e11a71b","Type":"ContainerStarted","Data":"628c4c49c842c68ad5f304513d44390ec6eb4d20f3b5b3e7081636fd7f5f6dc6"} Dec 03 17:38:08 crc kubenswrapper[4787]: I1203 17:38:08.667742 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ab2628-3d14-4c54-8314-1c395e11a71b","Type":"ContainerStarted","Data":"22406389f3496cc2d4b06fca93b46653743381ed5b2651875ee1ace56e44cef1"} Dec 03 17:38:10 crc kubenswrapper[4787]: I1203 17:38:10.689065 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ab2628-3d14-4c54-8314-1c395e11a71b","Type":"ContainerStarted","Data":"23a7f98b55503d768f3798cac1d6725fcc65cc412377f41760e4e386b3122828"} Dec 03 17:38:10 crc kubenswrapper[4787]: I1203 17:38:10.689630 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:38:10 crc kubenswrapper[4787]: I1203 17:38:10.723728 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.777114874 podStartE2EDuration="6.723706481s" podCreationTimestamp="2025-12-03 17:38:04 +0000 UTC" firstStartedPulling="2025-12-03 17:38:05.570958313 +0000 UTC m=+1502.388429272" lastFinishedPulling="2025-12-03 17:38:09.51754992 +0000 UTC m=+1506.335020879" observedRunningTime="2025-12-03 17:38:10.707813661 +0000 UTC m=+1507.525284630" watchObservedRunningTime="2025-12-03 17:38:10.723706481 +0000 UTC m=+1507.541177430" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.020806 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.751806 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-zq27q"] Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.753073 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.755779 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.755965 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.778817 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zq27q"] Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.809576 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pls7z\" (UniqueName: \"kubernetes.io/projected/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-kube-api-access-pls7z\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.809746 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-config-data\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.809894 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.809935 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-scripts\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.912193 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-config-data\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.912499 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.912520 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-scripts\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.912657 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pls7z\" (UniqueName: \"kubernetes.io/projected/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-kube-api-access-pls7z\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.918731 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-config-data\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.923033 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.932206 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-create-rb7vv"] Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.933615 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-rb7vv" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.941792 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-scripts\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.954195 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pls7z\" (UniqueName: \"kubernetes.io/projected/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-kube-api-access-pls7z\") pod \"nova-cell0-cell-mapping-zq27q\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.971580 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.974035 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:38:12 crc kubenswrapper[4787]: I1203 17:38:12.976444 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.015681 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-config-data\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.015823 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wwfwk\" (UniqueName: \"kubernetes.io/projected/6a476357-691f-402d-b70e-f81a8f37b27e-kube-api-access-wwfwk\") pod \"aodh-db-create-rb7vv\" (UID: \"6a476357-691f-402d-b70e-f81a8f37b27e\") " pod="openstack/aodh-db-create-rb7vv" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.015905 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a476357-691f-402d-b70e-f81a8f37b27e-operator-scripts\") pod \"aodh-db-create-rb7vv\" (UID: \"6a476357-691f-402d-b70e-f81a8f37b27e\") " pod="openstack/aodh-db-create-rb7vv" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.015944 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/752f318e-8037-4920-82cd-21cbd52994b6-logs\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.016040 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzt4t\" (UniqueName: \"kubernetes.io/projected/752f318e-8037-4920-82cd-21cbd52994b6-kube-api-access-zzt4t\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.016085 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.023807 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-rb7vv"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.062096 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.076704 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.108068 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.124811 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.127605 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.130673 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzt4t\" (UniqueName: \"kubernetes.io/projected/752f318e-8037-4920-82cd-21cbd52994b6-kube-api-access-zzt4t\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.131313 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.131396 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-config-data\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.131529 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wwfwk\" (UniqueName: \"kubernetes.io/projected/6a476357-691f-402d-b70e-f81a8f37b27e-kube-api-access-wwfwk\") pod \"aodh-db-create-rb7vv\" (UID: \"6a476357-691f-402d-b70e-f81a8f37b27e\") " pod="openstack/aodh-db-create-rb7vv" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.131689 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a476357-691f-402d-b70e-f81a8f37b27e-operator-scripts\") pod \"aodh-db-create-rb7vv\" (UID: \"6a476357-691f-402d-b70e-f81a8f37b27e\") " pod="openstack/aodh-db-create-rb7vv" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.131750 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/752f318e-8037-4920-82cd-21cbd52994b6-logs\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.132497 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/752f318e-8037-4920-82cd-21cbd52994b6-logs\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.136286 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a476357-691f-402d-b70e-f81a8f37b27e-operator-scripts\") pod \"aodh-db-create-rb7vv\" (UID: \"6a476357-691f-402d-b70e-f81a8f37b27e\") " pod="openstack/aodh-db-create-rb7vv" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.147894 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-config-data\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.166355 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wwfwk\" (UniqueName: \"kubernetes.io/projected/6a476357-691f-402d-b70e-f81a8f37b27e-kube-api-access-wwfwk\") pod \"aodh-db-create-rb7vv\" (UID: \"6a476357-691f-402d-b70e-f81a8f37b27e\") " pod="openstack/aodh-db-create-rb7vv" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.169905 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.174900 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzt4t\" (UniqueName: \"kubernetes.io/projected/752f318e-8037-4920-82cd-21cbd52994b6-kube-api-access-zzt4t\") pod \"nova-api-0\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.253905 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.254871 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9jkn\" (UniqueName: \"kubernetes.io/projected/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-kube-api-access-v9jkn\") pod \"nova-scheduler-0\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.255146 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-config-data\") pod \"nova-scheduler-0\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.322696 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.327947 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.338863 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.368474 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.368534 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9jkn\" (UniqueName: \"kubernetes.io/projected/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-kube-api-access-v9jkn\") pod \"nova-scheduler-0\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.368576 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-config-data\") pod \"nova-scheduler-0\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.399933 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-rb7vv" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.418227 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0c2d-account-create-update-5kwbn"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.430609 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0c2d-account-create-update-5kwbn" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.441681 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-db-secret" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.447331 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.472225 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.473088 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.473133 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4a7a56f-da6c-41e4-8147-ab175d3a0285-operator-scripts\") pod \"aodh-0c2d-account-create-update-5kwbn\" (UID: \"b4a7a56f-da6c-41e4-8147-ab175d3a0285\") " pod="openstack/aodh-0c2d-account-create-update-5kwbn" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.473212 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjgz8\" (UniqueName: \"kubernetes.io/projected/b96e978f-7e73-45ea-9cdc-73b4744bee47-kube-api-access-cjgz8\") pod \"nova-cell1-novncproxy-0\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.473314 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gzksc\" (UniqueName: \"kubernetes.io/projected/b4a7a56f-da6c-41e4-8147-ab175d3a0285-kube-api-access-gzksc\") pod \"aodh-0c2d-account-create-update-5kwbn\" (UID: \"b4a7a56f-da6c-41e4-8147-ab175d3a0285\") " pod="openstack/aodh-0c2d-account-create-update-5kwbn" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.486170 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.491724 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-config-data\") pod \"nova-scheduler-0\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.521203 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9jkn\" (UniqueName: \"kubernetes.io/projected/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-kube-api-access-v9jkn\") pod \"nova-scheduler-0\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.521682 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.541677 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.550973 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0c2d-account-create-update-5kwbn"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.573159 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.576329 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.576367 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4a7a56f-da6c-41e4-8147-ab175d3a0285-operator-scripts\") pod \"aodh-0c2d-account-create-update-5kwbn\" (UID: \"b4a7a56f-da6c-41e4-8147-ab175d3a0285\") " pod="openstack/aodh-0c2d-account-create-update-5kwbn" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.576426 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjgz8\" (UniqueName: \"kubernetes.io/projected/b96e978f-7e73-45ea-9cdc-73b4744bee47-kube-api-access-cjgz8\") pod \"nova-cell1-novncproxy-0\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.576475 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gzksc\" (UniqueName: \"kubernetes.io/projected/b4a7a56f-da6c-41e4-8147-ab175d3a0285-kube-api-access-gzksc\") pod \"aodh-0c2d-account-create-update-5kwbn\" (UID: \"b4a7a56f-da6c-41e4-8147-ab175d3a0285\") " pod="openstack/aodh-0c2d-account-create-update-5kwbn" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.576546 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.578667 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4a7a56f-da6c-41e4-8147-ab175d3a0285-operator-scripts\") pod \"aodh-0c2d-account-create-update-5kwbn\" (UID: \"b4a7a56f-da6c-41e4-8147-ab175d3a0285\") " pod="openstack/aodh-0c2d-account-create-update-5kwbn" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.581806 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.587559 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.610985 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjgz8\" (UniqueName: \"kubernetes.io/projected/b96e978f-7e73-45ea-9cdc-73b4744bee47-kube-api-access-cjgz8\") pod \"nova-cell1-novncproxy-0\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.626562 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gzksc\" (UniqueName: \"kubernetes.io/projected/b4a7a56f-da6c-41e4-8147-ab175d3a0285-kube-api-access-gzksc\") pod \"aodh-0c2d-account-create-update-5kwbn\" (UID: \"b4a7a56f-da6c-41e4-8147-ab175d3a0285\") " pod="openstack/aodh-0c2d-account-create-update-5kwbn" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.664452 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.672559 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.681569 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.737624 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.798972 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-config-data\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.799875 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4pfpt\" (UniqueName: \"kubernetes.io/projected/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-kube-api-access-4pfpt\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.799950 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.799972 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-logs\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.813050 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6w5hx"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.820270 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.838811 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.860401 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0c2d-account-create-update-5kwbn" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.873802 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6w5hx"] Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.902446 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.902494 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.902587 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.902620 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-logs\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.902742 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.902792 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-config\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.902893 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-svc\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.902989 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcq9p\" (UniqueName: \"kubernetes.io/projected/c19a0911-5c3e-4c36-96ec-064cd73b8bde-kube-api-access-qcq9p\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.903129 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-config-data\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.903229 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4pfpt\" (UniqueName: \"kubernetes.io/projected/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-kube-api-access-4pfpt\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.905564 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-logs\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.915637 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-config-data\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.915961 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:13 crc kubenswrapper[4787]: I1203 17:38:13.937408 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4pfpt\" (UniqueName: \"kubernetes.io/projected/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-kube-api-access-4pfpt\") pod \"nova-metadata-0\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " pod="openstack/nova-metadata-0" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.012713 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.014066 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.014118 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-config\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.014177 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-svc\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.014226 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcq9p\" (UniqueName: \"kubernetes.io/projected/c19a0911-5c3e-4c36-96ec-064cd73b8bde-kube-api-access-qcq9p\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.014313 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.014333 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.015240 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-sb\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.015855 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-nb\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.019541 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-svc\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.020872 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-swift-storage-0\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.034203 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-config\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.071721 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcq9p\" (UniqueName: \"kubernetes.io/projected/c19a0911-5c3e-4c36-96ec-064cd73b8bde-kube-api-access-qcq9p\") pod \"dnsmasq-dns-9b86998b5-6w5hx\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.077616 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-zq27q"] Dec 03 17:38:14 crc kubenswrapper[4787]: W1203 17:38:14.108141 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda5f00a92_19b0_4f7b_a3b8_66b3e720fab3.slice/crio-69e1e811cc07c583874d97af328f534805b3dafbdf0bf4644a0812f8825b18f6 WatchSource:0}: Error finding container 69e1e811cc07c583874d97af328f534805b3dafbdf0bf4644a0812f8825b18f6: Status 404 returned error can't find the container with id 69e1e811cc07c583874d97af328f534805b3dafbdf0bf4644a0812f8825b18f6 Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.161981 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.304571 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-create-rb7vv"] Dec 03 17:38:14 crc kubenswrapper[4787]: W1203 17:38:14.338591 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a476357_691f_402d_b70e_f81a8f37b27e.slice/crio-dfa14c6816a9f823cfecf4f53cd9501e468fb791a7cc8d5e394112327bab14e0 WatchSource:0}: Error finding container dfa14c6816a9f823cfecf4f53cd9501e468fb791a7cc8d5e394112327bab14e0: Status 404 returned error can't find the container with id dfa14c6816a9f823cfecf4f53cd9501e468fb791a7cc8d5e394112327bab14e0 Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.620728 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:38:14 crc kubenswrapper[4787]: W1203 17:38:14.675955 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7b772ec2_ecb8_4c2c_8cff_a7aef1a65339.slice/crio-cf7d787e7c0464886741ac0a68896bbae42670876ce7d94d10988938763b074b WatchSource:0}: Error finding container cf7d787e7c0464886741ac0a68896bbae42670876ce7d94d10988938763b074b: Status 404 returned error can't find the container with id cf7d787e7c0464886741ac0a68896bbae42670876ce7d94d10988938763b074b Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.750399 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339","Type":"ContainerStarted","Data":"cf7d787e7c0464886741ac0a68896bbae42670876ce7d94d10988938763b074b"} Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.761566 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zq27q" event={"ID":"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3","Type":"ContainerStarted","Data":"69e1e811cc07c583874d97af328f534805b3dafbdf0bf4644a0812f8825b18f6"} Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.770398 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-rb7vv" event={"ID":"6a476357-691f-402d-b70e-f81a8f37b27e","Type":"ContainerStarted","Data":"dfa14c6816a9f823cfecf4f53cd9501e468fb791a7cc8d5e394112327bab14e0"} Dec 03 17:38:14 crc kubenswrapper[4787]: I1203 17:38:14.783919 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.100294 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.129154 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0c2d-account-create-update-5kwbn"] Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.163637 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:15 crc kubenswrapper[4787]: W1203 17:38:15.175485 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc19a0911_5c3e_4c36_96ec_064cd73b8bde.slice/crio-aed22493ce8c642aa3c7d98467024e967fa4209bbc200661d6d2c7dad904540f WatchSource:0}: Error finding container aed22493ce8c642aa3c7d98467024e967fa4209bbc200661d6d2c7dad904540f: Status 404 returned error can't find the container with id aed22493ce8c642aa3c7d98467024e967fa4209bbc200661d6d2c7dad904540f Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.184379 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6w5hx"] Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.838666 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0c2d-account-create-update-5kwbn" event={"ID":"b4a7a56f-da6c-41e4-8147-ab175d3a0285","Type":"ContainerStarted","Data":"7e22aace230e3e0bfc95622f71791346bed85b1db35d4b468dfc47fc0c663c9f"} Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.845469 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zq27q" event={"ID":"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3","Type":"ContainerStarted","Data":"1d5521e67009ea94ea1bdda51184a2855b1c1209b56b5730ab6af30fca5119e6"} Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.850534 4787 generic.go:334] "Generic (PLEG): container finished" podID="6a476357-691f-402d-b70e-f81a8f37b27e" containerID="20cf456e5428a9bed8dec9ce1a32971183064f80563c71f03e150f7a0cb67760" exitCode=0 Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.850739 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-rb7vv" event={"ID":"6a476357-691f-402d-b70e-f81a8f37b27e","Type":"ContainerDied","Data":"20cf456e5428a9bed8dec9ce1a32971183064f80563c71f03e150f7a0cb67760"} Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.853686 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" event={"ID":"c19a0911-5c3e-4c36-96ec-064cd73b8bde","Type":"ContainerStarted","Data":"aed22493ce8c642aa3c7d98467024e967fa4209bbc200661d6d2c7dad904540f"} Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.861377 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b96e978f-7e73-45ea-9cdc-73b4744bee47","Type":"ContainerStarted","Data":"3a159799acec866dc5792a10af3e986a3e6cc8ae6b545e508296c59852983ff8"} Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.870207 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"66c05e7a-4cd6-4998-b86e-33cf0745e3b8","Type":"ContainerStarted","Data":"00fb75ce3858e2bd9deeaa24096f59076747bad3024276000d53df139203cc08"} Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.880861 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-zq27q" podStartSLOduration=3.880831937 podStartE2EDuration="3.880831937s" podCreationTimestamp="2025-12-03 17:38:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:15.872283566 +0000 UTC m=+1512.689754535" watchObservedRunningTime="2025-12-03 17:38:15.880831937 +0000 UTC m=+1512.698302906" Dec 03 17:38:15 crc kubenswrapper[4787]: I1203 17:38:15.883011 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"752f318e-8037-4920-82cd-21cbd52994b6","Type":"ContainerStarted","Data":"c8d136cb25745799cf03c09be714d4ca737d307d54b021073941f2cdff9caa29"} Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.016858 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-45bq2"] Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.019401 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.029364 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.029461 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.034411 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-45bq2"] Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.097124 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.097429 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kbvg\" (UniqueName: \"kubernetes.io/projected/e43d4111-655b-4843-bdc1-1465d8c5a050-kube-api-access-9kbvg\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.097603 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-scripts\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.097762 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-config-data\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.201995 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.202275 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kbvg\" (UniqueName: \"kubernetes.io/projected/e43d4111-655b-4843-bdc1-1465d8c5a050-kube-api-access-9kbvg\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.202378 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-scripts\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.202498 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-config-data\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.215970 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.218852 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-scripts\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.230455 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kbvg\" (UniqueName: \"kubernetes.io/projected/e43d4111-655b-4843-bdc1-1465d8c5a050-kube-api-access-9kbvg\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.235543 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-config-data\") pod \"nova-cell1-conductor-db-sync-45bq2\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.362718 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.899555 4787 generic.go:334] "Generic (PLEG): container finished" podID="b4a7a56f-da6c-41e4-8147-ab175d3a0285" containerID="0f939d4c6f7e73d503b443ea7afeeb178113b0790db90fe2bbf9eb309a6948da" exitCode=0 Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.899756 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0c2d-account-create-update-5kwbn" event={"ID":"b4a7a56f-da6c-41e4-8147-ab175d3a0285","Type":"ContainerDied","Data":"0f939d4c6f7e73d503b443ea7afeeb178113b0790db90fe2bbf9eb309a6948da"} Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.905101 4787 generic.go:334] "Generic (PLEG): container finished" podID="c19a0911-5c3e-4c36-96ec-064cd73b8bde" containerID="db4c2ec046c25c551a48c23635c66b7b80e0d6a76dad5ada71323359d4355caf" exitCode=0 Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.906168 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" event={"ID":"c19a0911-5c3e-4c36-96ec-064cd73b8bde","Type":"ContainerDied","Data":"db4c2ec046c25c551a48c23635c66b7b80e0d6a76dad5ada71323359d4355caf"} Dec 03 17:38:16 crc kubenswrapper[4787]: I1203 17:38:16.992794 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-45bq2"] Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.010011 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.013434 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="ceilometer-central-agent" containerID="cri-o://110cb8edea0aaf7f5dd3511f44ed88af2276c4b72f984eaadec63eec6188a897" gracePeriod=30 Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.014010 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="proxy-httpd" containerID="cri-o://23a7f98b55503d768f3798cac1d6725fcc65cc412377f41760e4e386b3122828" gracePeriod=30 Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.014313 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="ceilometer-notification-agent" containerID="cri-o://628c4c49c842c68ad5f304513d44390ec6eb4d20f3b5b3e7081636fd7f5f6dc6" gracePeriod=30 Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.014370 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="sg-core" containerID="cri-o://22406389f3496cc2d4b06fca93b46653743381ed5b2651875ee1ace56e44cef1" gracePeriod=30 Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.448400 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.484130 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:38:17 crc kubenswrapper[4787]: W1203 17:38:17.691620 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode43d4111_655b_4843_bdc1_1465d8c5a050.slice/crio-f1b50ca9a7e0b1da415f88e6e67d20e9ccfadc216f74b37be46da321b23d8e88 WatchSource:0}: Error finding container f1b50ca9a7e0b1da415f88e6e67d20e9ccfadc216f74b37be46da321b23d8e88: Status 404 returned error can't find the container with id f1b50ca9a7e0b1da415f88e6e67d20e9ccfadc216f74b37be46da321b23d8e88 Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.821112 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.821889 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell0-conductor-0" podUID="ac1922ce-f523-4e05-b9c1-676df04be85d" containerName="nova-cell0-conductor-conductor" containerID="cri-o://c74d5f2accf20704716e0459f555a75c24a58804dbcda99587a7d7a1f598b8d4" gracePeriod=30 Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.843687 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.891012 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.938239 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-45bq2" event={"ID":"e43d4111-655b-4843-bdc1-1465d8c5a050","Type":"ContainerStarted","Data":"f1b50ca9a7e0b1da415f88e6e67d20e9ccfadc216f74b37be46da321b23d8e88"} Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.959669 4787 generic.go:334] "Generic (PLEG): container finished" podID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerID="23a7f98b55503d768f3798cac1d6725fcc65cc412377f41760e4e386b3122828" exitCode=0 Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.959695 4787 generic.go:334] "Generic (PLEG): container finished" podID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerID="22406389f3496cc2d4b06fca93b46653743381ed5b2651875ee1ace56e44cef1" exitCode=2 Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.959704 4787 generic.go:334] "Generic (PLEG): container finished" podID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerID="110cb8edea0aaf7f5dd3511f44ed88af2276c4b72f984eaadec63eec6188a897" exitCode=0 Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.959894 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ab2628-3d14-4c54-8314-1c395e11a71b","Type":"ContainerDied","Data":"23a7f98b55503d768f3798cac1d6725fcc65cc412377f41760e4e386b3122828"} Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.959923 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ab2628-3d14-4c54-8314-1c395e11a71b","Type":"ContainerDied","Data":"22406389f3496cc2d4b06fca93b46653743381ed5b2651875ee1ace56e44cef1"} Dec 03 17:38:17 crc kubenswrapper[4787]: I1203 17:38:17.959932 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ab2628-3d14-4c54-8314-1c395e11a71b","Type":"ContainerDied","Data":"110cb8edea0aaf7f5dd3511f44ed88af2276c4b72f984eaadec63eec6188a897"} Dec 03 17:38:18 crc kubenswrapper[4787]: I1203 17:38:18.972735 4787 generic.go:334] "Generic (PLEG): container finished" podID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerID="628c4c49c842c68ad5f304513d44390ec6eb4d20f3b5b3e7081636fd7f5f6dc6" exitCode=0 Dec 03 17:38:18 crc kubenswrapper[4787]: I1203 17:38:18.972909 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ab2628-3d14-4c54-8314-1c395e11a71b","Type":"ContainerDied","Data":"628c4c49c842c68ad5f304513d44390ec6eb4d20f3b5b3e7081636fd7f5f6dc6"} Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.170593 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-rb7vv" Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.336953 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wwfwk\" (UniqueName: \"kubernetes.io/projected/6a476357-691f-402d-b70e-f81a8f37b27e-kube-api-access-wwfwk\") pod \"6a476357-691f-402d-b70e-f81a8f37b27e\" (UID: \"6a476357-691f-402d-b70e-f81a8f37b27e\") " Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.337274 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a476357-691f-402d-b70e-f81a8f37b27e-operator-scripts\") pod \"6a476357-691f-402d-b70e-f81a8f37b27e\" (UID: \"6a476357-691f-402d-b70e-f81a8f37b27e\") " Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.337908 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6a476357-691f-402d-b70e-f81a8f37b27e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6a476357-691f-402d-b70e-f81a8f37b27e" (UID: "6a476357-691f-402d-b70e-f81a8f37b27e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.338545 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6a476357-691f-402d-b70e-f81a8f37b27e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.346280 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a476357-691f-402d-b70e-f81a8f37b27e-kube-api-access-wwfwk" (OuterVolumeSpecName: "kube-api-access-wwfwk") pod "6a476357-691f-402d-b70e-f81a8f37b27e" (UID: "6a476357-691f-402d-b70e-f81a8f37b27e"). InnerVolumeSpecName "kube-api-access-wwfwk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.440157 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wwfwk\" (UniqueName: \"kubernetes.io/projected/6a476357-691f-402d-b70e-f81a8f37b27e-kube-api-access-wwfwk\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.704584 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0c2d-account-create-update-5kwbn" Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.851407 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gzksc\" (UniqueName: \"kubernetes.io/projected/b4a7a56f-da6c-41e4-8147-ab175d3a0285-kube-api-access-gzksc\") pod \"b4a7a56f-da6c-41e4-8147-ab175d3a0285\" (UID: \"b4a7a56f-da6c-41e4-8147-ab175d3a0285\") " Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.851555 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4a7a56f-da6c-41e4-8147-ab175d3a0285-operator-scripts\") pod \"b4a7a56f-da6c-41e4-8147-ab175d3a0285\" (UID: \"b4a7a56f-da6c-41e4-8147-ab175d3a0285\") " Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.852125 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b4a7a56f-da6c-41e4-8147-ab175d3a0285-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b4a7a56f-da6c-41e4-8147-ab175d3a0285" (UID: "b4a7a56f-da6c-41e4-8147-ab175d3a0285"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.853543 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b4a7a56f-da6c-41e4-8147-ab175d3a0285-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.872659 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4a7a56f-da6c-41e4-8147-ab175d3a0285-kube-api-access-gzksc" (OuterVolumeSpecName: "kube-api-access-gzksc") pod "b4a7a56f-da6c-41e4-8147-ab175d3a0285" (UID: "b4a7a56f-da6c-41e4-8147-ab175d3a0285"). InnerVolumeSpecName "kube-api-access-gzksc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:19 crc kubenswrapper[4787]: I1203 17:38:19.956055 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gzksc\" (UniqueName: \"kubernetes.io/projected/b4a7a56f-da6c-41e4-8147-ab175d3a0285-kube-api-access-gzksc\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.011047 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.025466 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-45bq2" event={"ID":"e43d4111-655b-4843-bdc1-1465d8c5a050","Type":"ContainerStarted","Data":"044d86c2a2fe3dbd8a3c6756e51f420f54addb310c01feaa7100fba82c302de5"} Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.054899 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0c2d-account-create-update-5kwbn" event={"ID":"b4a7a56f-da6c-41e4-8147-ab175d3a0285","Type":"ContainerDied","Data":"7e22aace230e3e0bfc95622f71791346bed85b1db35d4b468dfc47fc0c663c9f"} Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.054952 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e22aace230e3e0bfc95622f71791346bed85b1db35d4b468dfc47fc0c663c9f" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.055058 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0c2d-account-create-update-5kwbn" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.069676 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-create-rb7vv" event={"ID":"6a476357-691f-402d-b70e-f81a8f37b27e","Type":"ContainerDied","Data":"dfa14c6816a9f823cfecf4f53cd9501e468fb791a7cc8d5e394112327bab14e0"} Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.069867 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfa14c6816a9f823cfecf4f53cd9501e468fb791a7cc8d5e394112327bab14e0" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.070060 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-create-rb7vv" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.078282 4787 generic.go:334] "Generic (PLEG): container finished" podID="ac1922ce-f523-4e05-b9c1-676df04be85d" containerID="c74d5f2accf20704716e0459f555a75c24a58804dbcda99587a7d7a1f598b8d4" exitCode=0 Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.078361 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ac1922ce-f523-4e05-b9c1-676df04be85d","Type":"ContainerDied","Data":"c74d5f2accf20704716e0459f555a75c24a58804dbcda99587a7d7a1f598b8d4"} Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.080124 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7b772ec2-ecb8-4c2c-8cff-a7aef1a65339" containerName="nova-scheduler-scheduler" containerID="cri-o://174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271" gracePeriod=30 Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.104926 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-45bq2" podStartSLOduration=5.104903046 podStartE2EDuration="5.104903046s" podCreationTimestamp="2025-12-03 17:38:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:20.051913371 +0000 UTC m=+1516.869384330" watchObservedRunningTime="2025-12-03 17:38:20.104903046 +0000 UTC m=+1516.922374005" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.115940 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.138539884 podStartE2EDuration="8.115921264s" podCreationTimestamp="2025-12-03 17:38:12 +0000 UTC" firstStartedPulling="2025-12-03 17:38:14.697951187 +0000 UTC m=+1511.515422146" lastFinishedPulling="2025-12-03 17:38:19.675332567 +0000 UTC m=+1516.492803526" observedRunningTime="2025-12-03 17:38:20.093635491 +0000 UTC m=+1516.911106450" watchObservedRunningTime="2025-12-03 17:38:20.115921264 +0000 UTC m=+1516.933392223" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.171155 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-scripts\") pod \"85ab2628-3d14-4c54-8314-1c395e11a71b\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.171292 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2djxs\" (UniqueName: \"kubernetes.io/projected/85ab2628-3d14-4c54-8314-1c395e11a71b-kube-api-access-2djxs\") pod \"85ab2628-3d14-4c54-8314-1c395e11a71b\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.171383 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-log-httpd\") pod \"85ab2628-3d14-4c54-8314-1c395e11a71b\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.171427 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-sg-core-conf-yaml\") pod \"85ab2628-3d14-4c54-8314-1c395e11a71b\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.171498 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-run-httpd\") pod \"85ab2628-3d14-4c54-8314-1c395e11a71b\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.171568 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-combined-ca-bundle\") pod \"85ab2628-3d14-4c54-8314-1c395e11a71b\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.171602 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-config-data\") pod \"85ab2628-3d14-4c54-8314-1c395e11a71b\" (UID: \"85ab2628-3d14-4c54-8314-1c395e11a71b\") " Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.176478 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "85ab2628-3d14-4c54-8314-1c395e11a71b" (UID: "85ab2628-3d14-4c54-8314-1c395e11a71b"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.180265 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85ab2628-3d14-4c54-8314-1c395e11a71b-kube-api-access-2djxs" (OuterVolumeSpecName: "kube-api-access-2djxs") pod "85ab2628-3d14-4c54-8314-1c395e11a71b" (UID: "85ab2628-3d14-4c54-8314-1c395e11a71b"). InnerVolumeSpecName "kube-api-access-2djxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.180661 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "85ab2628-3d14-4c54-8314-1c395e11a71b" (UID: "85ab2628-3d14-4c54-8314-1c395e11a71b"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.186572 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-scripts" (OuterVolumeSpecName: "scripts") pod "85ab2628-3d14-4c54-8314-1c395e11a71b" (UID: "85ab2628-3d14-4c54-8314-1c395e11a71b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.265421 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "85ab2628-3d14-4c54-8314-1c395e11a71b" (UID: "85ab2628-3d14-4c54-8314-1c395e11a71b"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.276642 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.276683 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2djxs\" (UniqueName: \"kubernetes.io/projected/85ab2628-3d14-4c54-8314-1c395e11a71b-kube-api-access-2djxs\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.276698 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.276710 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.276722 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/85ab2628-3d14-4c54-8314-1c395e11a71b-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.308039 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85ab2628-3d14-4c54-8314-1c395e11a71b" (UID: "85ab2628-3d14-4c54-8314-1c395e11a71b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.359632 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-config-data" (OuterVolumeSpecName: "config-data") pod "85ab2628-3d14-4c54-8314-1c395e11a71b" (UID: "85ab2628-3d14-4c54-8314-1c395e11a71b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.380144 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.380184 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85ab2628-3d14-4c54-8314-1c395e11a71b-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.469457 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.583418 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-config-data\") pod \"ac1922ce-f523-4e05-b9c1-676df04be85d\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.583749 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pw7xw\" (UniqueName: \"kubernetes.io/projected/ac1922ce-f523-4e05-b9c1-676df04be85d-kube-api-access-pw7xw\") pod \"ac1922ce-f523-4e05-b9c1-676df04be85d\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.583980 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-combined-ca-bundle\") pod \"ac1922ce-f523-4e05-b9c1-676df04be85d\" (UID: \"ac1922ce-f523-4e05-b9c1-676df04be85d\") " Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.590390 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac1922ce-f523-4e05-b9c1-676df04be85d-kube-api-access-pw7xw" (OuterVolumeSpecName: "kube-api-access-pw7xw") pod "ac1922ce-f523-4e05-b9c1-676df04be85d" (UID: "ac1922ce-f523-4e05-b9c1-676df04be85d"). InnerVolumeSpecName "kube-api-access-pw7xw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.661647 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-config-data" (OuterVolumeSpecName: "config-data") pod "ac1922ce-f523-4e05-b9c1-676df04be85d" (UID: "ac1922ce-f523-4e05-b9c1-676df04be85d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.667956 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ac1922ce-f523-4e05-b9c1-676df04be85d" (UID: "ac1922ce-f523-4e05-b9c1-676df04be85d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.686374 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.686416 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pw7xw\" (UniqueName: \"kubernetes.io/projected/ac1922ce-f523-4e05-b9c1-676df04be85d-kube-api-access-pw7xw\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:20 crc kubenswrapper[4787]: I1203 17:38:20.686434 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ac1922ce-f523-4e05-b9c1-676df04be85d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.119579 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339","Type":"ContainerStarted","Data":"174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271"} Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.121933 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"752f318e-8037-4920-82cd-21cbd52994b6","Type":"ContainerStarted","Data":"7d1ac89bba90cfb50aa9c14eb344d6f682d27e8ce3a0b1a997ed10f52afc7566"} Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.121960 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"752f318e-8037-4920-82cd-21cbd52994b6","Type":"ContainerStarted","Data":"935b7ee11ea57c69e9a43e4a8ecf64f558e048c9598889a151a80fd84fc1589c"} Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.122110 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="752f318e-8037-4920-82cd-21cbd52994b6" containerName="nova-api-log" containerID="cri-o://935b7ee11ea57c69e9a43e4a8ecf64f558e048c9598889a151a80fd84fc1589c" gracePeriod=30 Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.122757 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="752f318e-8037-4920-82cd-21cbd52994b6" containerName="nova-api-api" containerID="cri-o://7d1ac89bba90cfb50aa9c14eb344d6f682d27e8ce3a0b1a997ed10f52afc7566" gracePeriod=30 Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.130401 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" event={"ID":"c19a0911-5c3e-4c36-96ec-064cd73b8bde","Type":"ContainerStarted","Data":"956351eadf8c23b4abdbbfb3c30e61772f4c35e4ebe96bdb403d10a529f50e2d"} Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.131264 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.134639 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"85ab2628-3d14-4c54-8314-1c395e11a71b","Type":"ContainerDied","Data":"73da631ec08fe842a9d99b12ac0262aec82089975a6faf43b40ae7ac69bdb046"} Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.134697 4787 scope.go:117] "RemoveContainer" containerID="23a7f98b55503d768f3798cac1d6725fcc65cc412377f41760e4e386b3122828" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.134871 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.139943 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b96e978f-7e73-45ea-9cdc-73b4744bee47","Type":"ContainerStarted","Data":"8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354"} Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.140220 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="b96e978f-7e73-45ea-9cdc-73b4744bee47" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354" gracePeriod=30 Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.148199 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"ac1922ce-f523-4e05-b9c1-676df04be85d","Type":"ContainerDied","Data":"93e2a4cbd36696d9e768224e4e1788161d246595e77ae2e97f6ae79a73d6401b"} Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.148311 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.153832 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=4.288005991 podStartE2EDuration="9.153814731s" podCreationTimestamp="2025-12-03 17:38:12 +0000 UTC" firstStartedPulling="2025-12-03 17:38:14.809519307 +0000 UTC m=+1511.626990266" lastFinishedPulling="2025-12-03 17:38:19.675328047 +0000 UTC m=+1516.492799006" observedRunningTime="2025-12-03 17:38:21.144778816 +0000 UTC m=+1517.962249775" watchObservedRunningTime="2025-12-03 17:38:21.153814731 +0000 UTC m=+1517.971285690" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.154873 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="66c05e7a-4cd6-4998-b86e-33cf0745e3b8" containerName="nova-metadata-log" containerID="cri-o://26cde0ac7cc50a73a5496a02f7d8aaeaa69bbfe4d01b060fda42ed94b3819f2e" gracePeriod=30 Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.155141 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="66c05e7a-4cd6-4998-b86e-33cf0745e3b8" containerName="nova-metadata-metadata" containerID="cri-o://557d08b6979e421789a5ed3b628eb957dea3461c3c6b3504a07d249d371b0fce" gracePeriod=30 Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.154920 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"66c05e7a-4cd6-4998-b86e-33cf0745e3b8","Type":"ContainerStarted","Data":"557d08b6979e421789a5ed3b628eb957dea3461c3c6b3504a07d249d371b0fce"} Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.156193 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"66c05e7a-4cd6-4998-b86e-33cf0745e3b8","Type":"ContainerStarted","Data":"26cde0ac7cc50a73a5496a02f7d8aaeaa69bbfe4d01b060fda42ed94b3819f2e"} Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.190448 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=3.581913598 podStartE2EDuration="8.190421532s" podCreationTimestamp="2025-12-03 17:38:13 +0000 UTC" firstStartedPulling="2025-12-03 17:38:15.131560355 +0000 UTC m=+1511.949031304" lastFinishedPulling="2025-12-03 17:38:19.740068279 +0000 UTC m=+1516.557539238" observedRunningTime="2025-12-03 17:38:21.166570126 +0000 UTC m=+1517.984041095" watchObservedRunningTime="2025-12-03 17:38:21.190421532 +0000 UTC m=+1518.007892491" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.195903 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" podStartSLOduration=8.19588931 podStartE2EDuration="8.19588931s" podCreationTimestamp="2025-12-03 17:38:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:21.185643192 +0000 UTC m=+1518.003114151" watchObservedRunningTime="2025-12-03 17:38:21.19588931 +0000 UTC m=+1518.013360269" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.210386 4787 scope.go:117] "RemoveContainer" containerID="22406389f3496cc2d4b06fca93b46653743381ed5b2651875ee1ace56e44cef1" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.242802 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.255066 4787 scope.go:117] "RemoveContainer" containerID="628c4c49c842c68ad5f304513d44390ec6eb4d20f3b5b3e7081636fd7f5f6dc6" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.255234 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.268839 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:38:21 crc kubenswrapper[4787]: E1203 17:38:21.269425 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="ceilometer-notification-agent" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269450 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="ceilometer-notification-agent" Dec 03 17:38:21 crc kubenswrapper[4787]: E1203 17:38:21.269488 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="sg-core" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269498 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="sg-core" Dec 03 17:38:21 crc kubenswrapper[4787]: E1203 17:38:21.269514 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac1922ce-f523-4e05-b9c1-676df04be85d" containerName="nova-cell0-conductor-conductor" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269523 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac1922ce-f523-4e05-b9c1-676df04be85d" containerName="nova-cell0-conductor-conductor" Dec 03 17:38:21 crc kubenswrapper[4787]: E1203 17:38:21.269534 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a476357-691f-402d-b70e-f81a8f37b27e" containerName="mariadb-database-create" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269542 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a476357-691f-402d-b70e-f81a8f37b27e" containerName="mariadb-database-create" Dec 03 17:38:21 crc kubenswrapper[4787]: E1203 17:38:21.269560 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="proxy-httpd" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269567 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="proxy-httpd" Dec 03 17:38:21 crc kubenswrapper[4787]: E1203 17:38:21.269586 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="ceilometer-central-agent" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269594 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="ceilometer-central-agent" Dec 03 17:38:21 crc kubenswrapper[4787]: E1203 17:38:21.269611 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4a7a56f-da6c-41e4-8147-ab175d3a0285" containerName="mariadb-account-create-update" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269620 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4a7a56f-da6c-41e4-8147-ab175d3a0285" containerName="mariadb-account-create-update" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269845 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="ceilometer-central-agent" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269889 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="proxy-httpd" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269905 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="ceilometer-notification-agent" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269923 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4a7a56f-da6c-41e4-8147-ab175d3a0285" containerName="mariadb-account-create-update" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269937 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" containerName="sg-core" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269959 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac1922ce-f523-4e05-b9c1-676df04be85d" containerName="nova-cell0-conductor-conductor" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.269968 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a476357-691f-402d-b70e-f81a8f37b27e" containerName="mariadb-database-create" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.270971 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.280332 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.282093 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.286702 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.6390359439999997 podStartE2EDuration="8.286685858s" podCreationTimestamp="2025-12-03 17:38:13 +0000 UTC" firstStartedPulling="2025-12-03 17:38:15.090699289 +0000 UTC m=+1511.908170248" lastFinishedPulling="2025-12-03 17:38:19.738349203 +0000 UTC m=+1516.555820162" observedRunningTime="2025-12-03 17:38:21.2553753 +0000 UTC m=+1518.072846279" watchObservedRunningTime="2025-12-03 17:38:21.286685858 +0000 UTC m=+1518.104156817" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.314620 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.328945 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.339100 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.343009 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.347387 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.347616 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.357597 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:21 crc kubenswrapper[4787]: E1203 17:38:21.382164 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod752f318e_8037_4920_82cd_21cbd52994b6.slice/crio-935b7ee11ea57c69e9a43e4a8ecf64f558e048c9598889a151a80fd84fc1589c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac1922ce_f523_4e05_b9c1_676df04be85d.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod66c05e7a_4cd6_4998_b86e_33cf0745e3b8.slice/crio-26cde0ac7cc50a73a5496a02f7d8aaeaa69bbfe4d01b060fda42ed94b3819f2e.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod752f318e_8037_4920_82cd_21cbd52994b6.slice/crio-conmon-935b7ee11ea57c69e9a43e4a8ecf64f558e048c9598889a151a80fd84fc1589c.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.398656 4787 scope.go:117] "RemoveContainer" containerID="110cb8edea0aaf7f5dd3511f44ed88af2276c4b72f984eaadec63eec6188a897" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.426619 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42612f1a-6592-480c-b475-865267ee2f9f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"42612f1a-6592-480c-b475-865267ee2f9f\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.426810 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.426842 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-run-httpd\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.426899 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5kl8\" (UniqueName: \"kubernetes.io/projected/758c0073-2001-4fa7-b1f1-e5350aae3e33-kube-api-access-p5kl8\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.426933 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-log-httpd\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.426970 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.427054 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-scripts\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.427098 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-config-data\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.427183 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42612f1a-6592-480c-b475-865267ee2f9f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"42612f1a-6592-480c-b475-865267ee2f9f\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.427225 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-87hff\" (UniqueName: \"kubernetes.io/projected/42612f1a-6592-480c-b475-865267ee2f9f-kube-api-access-87hff\") pod \"nova-cell0-conductor-0\" (UID: \"42612f1a-6592-480c-b475-865267ee2f9f\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.457390 4787 scope.go:117] "RemoveContainer" containerID="c74d5f2accf20704716e0459f555a75c24a58804dbcda99587a7d7a1f598b8d4" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.529399 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42612f1a-6592-480c-b475-865267ee2f9f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"42612f1a-6592-480c-b475-865267ee2f9f\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.529695 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.529794 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-run-httpd\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.529896 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5kl8\" (UniqueName: \"kubernetes.io/projected/758c0073-2001-4fa7-b1f1-e5350aae3e33-kube-api-access-p5kl8\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.529988 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-log-httpd\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.530115 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.530230 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-scripts\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.530320 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-config-data\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.530431 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42612f1a-6592-480c-b475-865267ee2f9f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"42612f1a-6592-480c-b475-865267ee2f9f\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.530588 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-log-httpd\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.530598 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-87hff\" (UniqueName: \"kubernetes.io/projected/42612f1a-6592-480c-b475-865267ee2f9f-kube-api-access-87hff\") pod \"nova-cell0-conductor-0\" (UID: \"42612f1a-6592-480c-b475-865267ee2f9f\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.530438 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-run-httpd\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.535820 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42612f1a-6592-480c-b475-865267ee2f9f-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"42612f1a-6592-480c-b475-865267ee2f9f\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.535915 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-scripts\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.536087 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.547279 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42612f1a-6592-480c-b475-865267ee2f9f-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"42612f1a-6592-480c-b475-865267ee2f9f\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.547506 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.548460 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-config-data\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.551970 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-87hff\" (UniqueName: \"kubernetes.io/projected/42612f1a-6592-480c-b475-865267ee2f9f-kube-api-access-87hff\") pod \"nova-cell0-conductor-0\" (UID: \"42612f1a-6592-480c-b475-865267ee2f9f\") " pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.563780 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5kl8\" (UniqueName: \"kubernetes.io/projected/758c0073-2001-4fa7-b1f1-e5350aae3e33-kube-api-access-p5kl8\") pod \"ceilometer-0\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.695920 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.712634 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.790640 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85ab2628-3d14-4c54-8314-1c395e11a71b" path="/var/lib/kubelet/pods/85ab2628-3d14-4c54-8314-1c395e11a71b/volumes" Dec 03 17:38:21 crc kubenswrapper[4787]: I1203 17:38:21.792732 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac1922ce-f523-4e05-b9c1-676df04be85d" path="/var/lib/kubelet/pods/ac1922ce-f523-4e05-b9c1-676df04be85d/volumes" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.196470 4787 generic.go:334] "Generic (PLEG): container finished" podID="752f318e-8037-4920-82cd-21cbd52994b6" containerID="7d1ac89bba90cfb50aa9c14eb344d6f682d27e8ce3a0b1a997ed10f52afc7566" exitCode=0 Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.196883 4787 generic.go:334] "Generic (PLEG): container finished" podID="752f318e-8037-4920-82cd-21cbd52994b6" containerID="935b7ee11ea57c69e9a43e4a8ecf64f558e048c9598889a151a80fd84fc1589c" exitCode=143 Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.196931 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"752f318e-8037-4920-82cd-21cbd52994b6","Type":"ContainerDied","Data":"7d1ac89bba90cfb50aa9c14eb344d6f682d27e8ce3a0b1a997ed10f52afc7566"} Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.196957 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"752f318e-8037-4920-82cd-21cbd52994b6","Type":"ContainerDied","Data":"935b7ee11ea57c69e9a43e4a8ecf64f558e048c9598889a151a80fd84fc1589c"} Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.214477 4787 generic.go:334] "Generic (PLEG): container finished" podID="66c05e7a-4cd6-4998-b86e-33cf0745e3b8" containerID="557d08b6979e421789a5ed3b628eb957dea3461c3c6b3504a07d249d371b0fce" exitCode=0 Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.214506 4787 generic.go:334] "Generic (PLEG): container finished" podID="66c05e7a-4cd6-4998-b86e-33cf0745e3b8" containerID="26cde0ac7cc50a73a5496a02f7d8aaeaa69bbfe4d01b060fda42ed94b3819f2e" exitCode=143 Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.215114 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"66c05e7a-4cd6-4998-b86e-33cf0745e3b8","Type":"ContainerDied","Data":"557d08b6979e421789a5ed3b628eb957dea3461c3c6b3504a07d249d371b0fce"} Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.215170 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"66c05e7a-4cd6-4998-b86e-33cf0745e3b8","Type":"ContainerDied","Data":"26cde0ac7cc50a73a5496a02f7d8aaeaa69bbfe4d01b060fda42ed94b3819f2e"} Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.312040 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.350796 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4pfpt\" (UniqueName: \"kubernetes.io/projected/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-kube-api-access-4pfpt\") pod \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.351035 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-logs\") pod \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.351105 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-config-data\") pod \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.351170 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-combined-ca-bundle\") pod \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\" (UID: \"66c05e7a-4cd6-4998-b86e-33cf0745e3b8\") " Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.351349 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-logs" (OuterVolumeSpecName: "logs") pod "66c05e7a-4cd6-4998-b86e-33cf0745e3b8" (UID: "66c05e7a-4cd6-4998-b86e-33cf0745e3b8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.351657 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.363683 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-kube-api-access-4pfpt" (OuterVolumeSpecName: "kube-api-access-4pfpt") pod "66c05e7a-4cd6-4998-b86e-33cf0745e3b8" (UID: "66c05e7a-4cd6-4998-b86e-33cf0745e3b8"). InnerVolumeSpecName "kube-api-access-4pfpt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.387566 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-config-data" (OuterVolumeSpecName: "config-data") pod "66c05e7a-4cd6-4998-b86e-33cf0745e3b8" (UID: "66c05e7a-4cd6-4998-b86e-33cf0745e3b8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.441456 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "66c05e7a-4cd6-4998-b86e-33cf0745e3b8" (UID: "66c05e7a-4cd6-4998-b86e-33cf0745e3b8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.453278 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.453320 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.453335 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4pfpt\" (UniqueName: \"kubernetes.io/projected/66c05e7a-4cd6-4998-b86e-33cf0745e3b8-kube-api-access-4pfpt\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.539940 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.555235 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzt4t\" (UniqueName: \"kubernetes.io/projected/752f318e-8037-4920-82cd-21cbd52994b6-kube-api-access-zzt4t\") pod \"752f318e-8037-4920-82cd-21cbd52994b6\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.555477 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-combined-ca-bundle\") pod \"752f318e-8037-4920-82cd-21cbd52994b6\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.555579 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-config-data\") pod \"752f318e-8037-4920-82cd-21cbd52994b6\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.555675 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/752f318e-8037-4920-82cd-21cbd52994b6-logs\") pod \"752f318e-8037-4920-82cd-21cbd52994b6\" (UID: \"752f318e-8037-4920-82cd-21cbd52994b6\") " Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.556537 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/752f318e-8037-4920-82cd-21cbd52994b6-logs" (OuterVolumeSpecName: "logs") pod "752f318e-8037-4920-82cd-21cbd52994b6" (UID: "752f318e-8037-4920-82cd-21cbd52994b6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.565354 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/752f318e-8037-4920-82cd-21cbd52994b6-kube-api-access-zzt4t" (OuterVolumeSpecName: "kube-api-access-zzt4t") pod "752f318e-8037-4920-82cd-21cbd52994b6" (UID: "752f318e-8037-4920-82cd-21cbd52994b6"). InnerVolumeSpecName "kube-api-access-zzt4t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.611694 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-config-data" (OuterVolumeSpecName: "config-data") pod "752f318e-8037-4920-82cd-21cbd52994b6" (UID: "752f318e-8037-4920-82cd-21cbd52994b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.612105 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "752f318e-8037-4920-82cd-21cbd52994b6" (UID: "752f318e-8037-4920-82cd-21cbd52994b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.667520 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.667561 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/752f318e-8037-4920-82cd-21cbd52994b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.667574 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/752f318e-8037-4920-82cd-21cbd52994b6-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.667584 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzt4t\" (UniqueName: \"kubernetes.io/projected/752f318e-8037-4920-82cd-21cbd52994b6-kube-api-access-zzt4t\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.743187 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:22 crc kubenswrapper[4787]: W1203 17:38:22.745724 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod758c0073_2001_4fa7_b1f1_e5350aae3e33.slice/crio-9f68b33ba2baf7d6659a5799041e1aa5772c9aaff312161bbff190bf4bbd784f WatchSource:0}: Error finding container 9f68b33ba2baf7d6659a5799041e1aa5772c9aaff312161bbff190bf4bbd784f: Status 404 returned error can't find the container with id 9f68b33ba2baf7d6659a5799041e1aa5772c9aaff312161bbff190bf4bbd784f Dec 03 17:38:22 crc kubenswrapper[4787]: W1203 17:38:22.746817 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod42612f1a_6592_480c_b475_865267ee2f9f.slice/crio-f57d055141a6f5ac07171604998aa5faac1bff39f42817bb73e61e13b2d7a7dd WatchSource:0}: Error finding container f57d055141a6f5ac07171604998aa5faac1bff39f42817bb73e61e13b2d7a7dd: Status 404 returned error can't find the container with id f57d055141a6f5ac07171604998aa5faac1bff39f42817bb73e61e13b2d7a7dd Dec 03 17:38:22 crc kubenswrapper[4787]: I1203 17:38:22.757367 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.229589 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"66c05e7a-4cd6-4998-b86e-33cf0745e3b8","Type":"ContainerDied","Data":"00fb75ce3858e2bd9deeaa24096f59076747bad3024276000d53df139203cc08"} Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.229925 4787 scope.go:117] "RemoveContainer" containerID="557d08b6979e421789a5ed3b628eb957dea3461c3c6b3504a07d249d371b0fce" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.229657 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.231041 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"42612f1a-6592-480c-b475-865267ee2f9f","Type":"ContainerStarted","Data":"f57d055141a6f5ac07171604998aa5faac1bff39f42817bb73e61e13b2d7a7dd"} Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.233119 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"758c0073-2001-4fa7-b1f1-e5350aae3e33","Type":"ContainerStarted","Data":"9f68b33ba2baf7d6659a5799041e1aa5772c9aaff312161bbff190bf4bbd784f"} Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.236473 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"752f318e-8037-4920-82cd-21cbd52994b6","Type":"ContainerDied","Data":"c8d136cb25745799cf03c09be714d4ca737d307d54b021073941f2cdff9caa29"} Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.236667 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.262517 4787 scope.go:117] "RemoveContainer" containerID="26cde0ac7cc50a73a5496a02f7d8aaeaa69bbfe4d01b060fda42ed94b3819f2e" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.274810 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.299337 4787 scope.go:117] "RemoveContainer" containerID="7d1ac89bba90cfb50aa9c14eb344d6f682d27e8ce3a0b1a997ed10f52afc7566" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.308231 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.333491 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.334443 4787 scope.go:117] "RemoveContainer" containerID="935b7ee11ea57c69e9a43e4a8ecf64f558e048c9598889a151a80fd84fc1589c" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.347125 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.372605 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:23 crc kubenswrapper[4787]: E1203 17:38:23.373194 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="752f318e-8037-4920-82cd-21cbd52994b6" containerName="nova-api-log" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.373211 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="752f318e-8037-4920-82cd-21cbd52994b6" containerName="nova-api-log" Dec 03 17:38:23 crc kubenswrapper[4787]: E1203 17:38:23.373247 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c05e7a-4cd6-4998-b86e-33cf0745e3b8" containerName="nova-metadata-log" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.373255 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c05e7a-4cd6-4998-b86e-33cf0745e3b8" containerName="nova-metadata-log" Dec 03 17:38:23 crc kubenswrapper[4787]: E1203 17:38:23.373281 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="752f318e-8037-4920-82cd-21cbd52994b6" containerName="nova-api-api" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.373289 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="752f318e-8037-4920-82cd-21cbd52994b6" containerName="nova-api-api" Dec 03 17:38:23 crc kubenswrapper[4787]: E1203 17:38:23.373315 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66c05e7a-4cd6-4998-b86e-33cf0745e3b8" containerName="nova-metadata-metadata" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.373322 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="66c05e7a-4cd6-4998-b86e-33cf0745e3b8" containerName="nova-metadata-metadata" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.373549 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="66c05e7a-4cd6-4998-b86e-33cf0745e3b8" containerName="nova-metadata-metadata" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.373566 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="752f318e-8037-4920-82cd-21cbd52994b6" containerName="nova-api-log" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.373582 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="752f318e-8037-4920-82cd-21cbd52994b6" containerName="nova-api-api" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.373602 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="66c05e7a-4cd6-4998-b86e-33cf0745e3b8" containerName="nova-metadata-log" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.375520 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.381828 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.382097 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.392340 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.411980 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.414185 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.416567 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.428610 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.500826 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-logs\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.500950 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-config-data\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.500970 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.501007 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.501256 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmzkg\" (UniqueName: \"kubernetes.io/projected/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-kube-api-access-zmzkg\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.541816 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-8jchz"] Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.543694 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.546656 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.546742 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.546846 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-q4n2s" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.548873 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.555110 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-8jchz"] Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.574266 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.603809 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-logs\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.603892 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-config-data\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.603949 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.604359 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-logs\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.605514 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.605604 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.605630 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85e3cc25-5c94-46d1-a107-a3258de53909-logs\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.605894 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmzkg\" (UniqueName: \"kubernetes.io/projected/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-kube-api-access-zmzkg\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.606043 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-config-data\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.606096 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jc4b2\" (UniqueName: \"kubernetes.io/projected/85e3cc25-5c94-46d1-a107-a3258de53909-kube-api-access-jc4b2\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.610983 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.613598 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-config-data\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.629078 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.635525 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmzkg\" (UniqueName: \"kubernetes.io/projected/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-kube-api-access-zmzkg\") pod \"nova-metadata-0\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.707669 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85e3cc25-5c94-46d1-a107-a3258de53909-logs\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.707714 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.707749 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-combined-ca-bundle\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.707800 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-scripts\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.707833 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-config-data\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.707854 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-config-data\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.707872 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jc4b2\" (UniqueName: \"kubernetes.io/projected/85e3cc25-5c94-46d1-a107-a3258de53909-kube-api-access-jc4b2\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.707906 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npxt9\" (UniqueName: \"kubernetes.io/projected/92bdbf6e-87dc-4a93-aefd-26b002566663-kube-api-access-npxt9\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.708410 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85e3cc25-5c94-46d1-a107-a3258de53909-logs\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.709755 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.711920 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.716519 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-config-data\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.736532 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jc4b2\" (UniqueName: \"kubernetes.io/projected/85e3cc25-5c94-46d1-a107-a3258de53909-kube-api-access-jc4b2\") pod \"nova-api-0\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.751639 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.794585 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66c05e7a-4cd6-4998-b86e-33cf0745e3b8" path="/var/lib/kubelet/pods/66c05e7a-4cd6-4998-b86e-33cf0745e3b8/volumes" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.795448 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="752f318e-8037-4920-82cd-21cbd52994b6" path="/var/lib/kubelet/pods/752f318e-8037-4920-82cd-21cbd52994b6/volumes" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.817593 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-combined-ca-bundle\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.817760 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-scripts\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.817910 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-config-data\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.818059 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npxt9\" (UniqueName: \"kubernetes.io/projected/92bdbf6e-87dc-4a93-aefd-26b002566663-kube-api-access-npxt9\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.830375 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-scripts\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.838745 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-config-data\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.840055 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.841227 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-combined-ca-bundle\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:23 crc kubenswrapper[4787]: I1203 17:38:23.873398 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npxt9\" (UniqueName: \"kubernetes.io/projected/92bdbf6e-87dc-4a93-aefd-26b002566663-kube-api-access-npxt9\") pod \"aodh-db-sync-8jchz\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:24 crc kubenswrapper[4787]: I1203 17:38:24.160916 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:24 crc kubenswrapper[4787]: I1203 17:38:24.321457 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"42612f1a-6592-480c-b475-865267ee2f9f","Type":"ContainerStarted","Data":"ae408affd337f17e30a45dc7bac95392fc50a82a155d88cba12c95597e459222"} Dec 03 17:38:24 crc kubenswrapper[4787]: I1203 17:38:24.323910 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:24 crc kubenswrapper[4787]: I1203 17:38:24.358796 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.35876746 podStartE2EDuration="3.35876746s" podCreationTimestamp="2025-12-03 17:38:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:24.343877007 +0000 UTC m=+1521.161347966" watchObservedRunningTime="2025-12-03 17:38:24.35876746 +0000 UTC m=+1521.176238419" Dec 03 17:38:24 crc kubenswrapper[4787]: W1203 17:38:24.451536 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod85e3cc25_5c94_46d1_a107_a3258de53909.slice/crio-4943edae8da058bde0b89216c43952f0bb34251e85735bff2ce9ef31e0bee853 WatchSource:0}: Error finding container 4943edae8da058bde0b89216c43952f0bb34251e85735bff2ce9ef31e0bee853: Status 404 returned error can't find the container with id 4943edae8da058bde0b89216c43952f0bb34251e85735bff2ce9ef31e0bee853 Dec 03 17:38:24 crc kubenswrapper[4787]: I1203 17:38:24.458000 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:24 crc kubenswrapper[4787]: I1203 17:38:24.786814 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:24 crc kubenswrapper[4787]: I1203 17:38:24.812241 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-8jchz"] Dec 03 17:38:25 crc kubenswrapper[4787]: I1203 17:38:25.341602 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85e3cc25-5c94-46d1-a107-a3258de53909","Type":"ContainerStarted","Data":"38642b74aa386c284e7955a39728d7306137754244f5832b47ecca3313e1b0f5"} Dec 03 17:38:25 crc kubenswrapper[4787]: I1203 17:38:25.342236 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85e3cc25-5c94-46d1-a107-a3258de53909","Type":"ContainerStarted","Data":"4943edae8da058bde0b89216c43952f0bb34251e85735bff2ce9ef31e0bee853"} Dec 03 17:38:25 crc kubenswrapper[4787]: I1203 17:38:25.344791 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82e8b07e-6b53-4ae7-8570-fa1dc92bd975","Type":"ContainerStarted","Data":"4691eb61c2989b562475460aedde64faf9f0efa1058ca5d3a29c3f9179e079b7"} Dec 03 17:38:25 crc kubenswrapper[4787]: I1203 17:38:25.347156 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-8jchz" event={"ID":"92bdbf6e-87dc-4a93-aefd-26b002566663","Type":"ContainerStarted","Data":"16ec847b2ca36ce1524408f49589720b3d3f4827e7e94138533ceb4c2a4dc841"} Dec 03 17:38:25 crc kubenswrapper[4787]: I1203 17:38:25.349492 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"758c0073-2001-4fa7-b1f1-e5350aae3e33","Type":"ContainerStarted","Data":"73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420"} Dec 03 17:38:26 crc kubenswrapper[4787]: I1203 17:38:26.375111 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85e3cc25-5c94-46d1-a107-a3258de53909","Type":"ContainerStarted","Data":"bee944dc494d18fd1543317d49ac4b87f18be6917f77605cd6e8485043777489"} Dec 03 17:38:26 crc kubenswrapper[4787]: I1203 17:38:26.378347 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82e8b07e-6b53-4ae7-8570-fa1dc92bd975","Type":"ContainerStarted","Data":"35fcd12546c3bc57ab7c66269b709ed60bab9892202b0be9bee36a2f23dd4843"} Dec 03 17:38:26 crc kubenswrapper[4787]: I1203 17:38:26.378384 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82e8b07e-6b53-4ae7-8570-fa1dc92bd975","Type":"ContainerStarted","Data":"1b3f121048d270d93b69e86c769d4f15d847505815a668ea1cab35e3cfeaa06f"} Dec 03 17:38:26 crc kubenswrapper[4787]: I1203 17:38:26.394059 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"758c0073-2001-4fa7-b1f1-e5350aae3e33","Type":"ContainerStarted","Data":"35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e"} Dec 03 17:38:26 crc kubenswrapper[4787]: I1203 17:38:26.396526 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.396515394 podStartE2EDuration="3.396515394s" podCreationTimestamp="2025-12-03 17:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:26.394719935 +0000 UTC m=+1523.212190894" watchObservedRunningTime="2025-12-03 17:38:26.396515394 +0000 UTC m=+1523.213986353" Dec 03 17:38:26 crc kubenswrapper[4787]: I1203 17:38:26.431199 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.431178472 podStartE2EDuration="3.431178472s" podCreationTimestamp="2025-12-03 17:38:23 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:26.424261055 +0000 UTC m=+1523.241732014" watchObservedRunningTime="2025-12-03 17:38:26.431178472 +0000 UTC m=+1523.248649431" Dec 03 17:38:27 crc kubenswrapper[4787]: I1203 17:38:27.419210 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"758c0073-2001-4fa7-b1f1-e5350aae3e33","Type":"ContainerStarted","Data":"d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d"} Dec 03 17:38:28 crc kubenswrapper[4787]: I1203 17:38:28.709984 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:38:28 crc kubenswrapper[4787]: I1203 17:38:28.710221 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:38:29 crc kubenswrapper[4787]: I1203 17:38:29.165224 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:38:29 crc kubenswrapper[4787]: I1203 17:38:29.241162 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jlk65"] Dec 03 17:38:29 crc kubenswrapper[4787]: I1203 17:38:29.241417 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" podUID="b67a4816-85d5-404a-a22d-6cfbc455d286" containerName="dnsmasq-dns" containerID="cri-o://b2c6b8eb60b0320b23d2a30b6c8301c807adc10713a22a67b0f04a7e88b4bbcf" gracePeriod=10 Dec 03 17:38:29 crc kubenswrapper[4787]: I1203 17:38:29.447252 4787 generic.go:334] "Generic (PLEG): container finished" podID="b67a4816-85d5-404a-a22d-6cfbc455d286" containerID="b2c6b8eb60b0320b23d2a30b6c8301c807adc10713a22a67b0f04a7e88b4bbcf" exitCode=0 Dec 03 17:38:29 crc kubenswrapper[4787]: I1203 17:38:29.447338 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" event={"ID":"b67a4816-85d5-404a-a22d-6cfbc455d286","Type":"ContainerDied","Data":"b2c6b8eb60b0320b23d2a30b6c8301c807adc10713a22a67b0f04a7e88b4bbcf"} Dec 03 17:38:29 crc kubenswrapper[4787]: I1203 17:38:29.448926 4787 generic.go:334] "Generic (PLEG): container finished" podID="a5f00a92-19b0-4f7b-a3b8-66b3e720fab3" containerID="1d5521e67009ea94ea1bdda51184a2855b1c1209b56b5730ab6af30fca5119e6" exitCode=0 Dec 03 17:38:29 crc kubenswrapper[4787]: I1203 17:38:29.448976 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zq27q" event={"ID":"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3","Type":"ContainerDied","Data":"1d5521e67009ea94ea1bdda51184a2855b1c1209b56b5730ab6af30fca5119e6"} Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.476150 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-zq27q" event={"ID":"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3","Type":"ContainerDied","Data":"69e1e811cc07c583874d97af328f534805b3dafbdf0bf4644a0812f8825b18f6"} Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.476776 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="69e1e811cc07c583874d97af328f534805b3dafbdf0bf4644a0812f8825b18f6" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.543630 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.612944 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-config-data\") pod \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.613571 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-combined-ca-bundle\") pod \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.613798 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-scripts\") pod \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.613927 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pls7z\" (UniqueName: \"kubernetes.io/projected/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-kube-api-access-pls7z\") pod \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\" (UID: \"a5f00a92-19b0-4f7b-a3b8-66b3e720fab3\") " Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.621335 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-kube-api-access-pls7z" (OuterVolumeSpecName: "kube-api-access-pls7z") pod "a5f00a92-19b0-4f7b-a3b8-66b3e720fab3" (UID: "a5f00a92-19b0-4f7b-a3b8-66b3e720fab3"). InnerVolumeSpecName "kube-api-access-pls7z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.621534 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-scripts" (OuterVolumeSpecName: "scripts") pod "a5f00a92-19b0-4f7b-a3b8-66b3e720fab3" (UID: "a5f00a92-19b0-4f7b-a3b8-66b3e720fab3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.670263 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-config-data" (OuterVolumeSpecName: "config-data") pod "a5f00a92-19b0-4f7b-a3b8-66b3e720fab3" (UID: "a5f00a92-19b0-4f7b-a3b8-66b3e720fab3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.680986 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a5f00a92-19b0-4f7b-a3b8-66b3e720fab3" (UID: "a5f00a92-19b0-4f7b-a3b8-66b3e720fab3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.737908 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.737979 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.737995 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.738049 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pls7z\" (UniqueName: \"kubernetes.io/projected/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3-kube-api-access-pls7z\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.744145 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.833246 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.942057 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-svc\") pod \"b67a4816-85d5-404a-a22d-6cfbc455d286\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.942216 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-swift-storage-0\") pod \"b67a4816-85d5-404a-a22d-6cfbc455d286\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.942306 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-config\") pod \"b67a4816-85d5-404a-a22d-6cfbc455d286\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.942363 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zp4hn\" (UniqueName: \"kubernetes.io/projected/b67a4816-85d5-404a-a22d-6cfbc455d286-kube-api-access-zp4hn\") pod \"b67a4816-85d5-404a-a22d-6cfbc455d286\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.942451 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-sb\") pod \"b67a4816-85d5-404a-a22d-6cfbc455d286\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.942662 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-nb\") pod \"b67a4816-85d5-404a-a22d-6cfbc455d286\" (UID: \"b67a4816-85d5-404a-a22d-6cfbc455d286\") " Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.948911 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b67a4816-85d5-404a-a22d-6cfbc455d286-kube-api-access-zp4hn" (OuterVolumeSpecName: "kube-api-access-zp4hn") pod "b67a4816-85d5-404a-a22d-6cfbc455d286" (UID: "b67a4816-85d5-404a-a22d-6cfbc455d286"). InnerVolumeSpecName "kube-api-access-zp4hn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.994776 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b67a4816-85d5-404a-a22d-6cfbc455d286" (UID: "b67a4816-85d5-404a-a22d-6cfbc455d286"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:38:31 crc kubenswrapper[4787]: I1203 17:38:31.995569 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b67a4816-85d5-404a-a22d-6cfbc455d286" (UID: "b67a4816-85d5-404a-a22d-6cfbc455d286"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.001853 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-config" (OuterVolumeSpecName: "config") pod "b67a4816-85d5-404a-a22d-6cfbc455d286" (UID: "b67a4816-85d5-404a-a22d-6cfbc455d286"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.014885 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b67a4816-85d5-404a-a22d-6cfbc455d286" (UID: "b67a4816-85d5-404a-a22d-6cfbc455d286"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.019121 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b67a4816-85d5-404a-a22d-6cfbc455d286" (UID: "b67a4816-85d5-404a-a22d-6cfbc455d286"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.046291 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zp4hn\" (UniqueName: \"kubernetes.io/projected/b67a4816-85d5-404a-a22d-6cfbc455d286-kube-api-access-zp4hn\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.046324 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.046335 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.046347 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.046357 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.046366 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b67a4816-85d5-404a-a22d-6cfbc455d286-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.489272 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" event={"ID":"b67a4816-85d5-404a-a22d-6cfbc455d286","Type":"ContainerDied","Data":"dd3f44d97441b04422c9b10ecc784e281e8bde8e7b954f2d5274a7972ddc9ef6"} Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.489630 4787 scope.go:117] "RemoveContainer" containerID="b2c6b8eb60b0320b23d2a30b6c8301c807adc10713a22a67b0f04a7e88b4bbcf" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.489352 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.492813 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-8jchz" event={"ID":"92bdbf6e-87dc-4a93-aefd-26b002566663","Type":"ContainerStarted","Data":"021fa014eff70f032132ca3763ca662441515e52086cce85320c771165ecb316"} Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.497057 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-zq27q" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.497091 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"758c0073-2001-4fa7-b1f1-e5350aae3e33","Type":"ContainerStarted","Data":"715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65"} Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.511463 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-8jchz" podStartSLOduration=2.241461556 podStartE2EDuration="9.511448849s" podCreationTimestamp="2025-12-03 17:38:23 +0000 UTC" firstStartedPulling="2025-12-03 17:38:24.817945831 +0000 UTC m=+1521.635416790" lastFinishedPulling="2025-12-03 17:38:32.087933124 +0000 UTC m=+1528.905404083" observedRunningTime="2025-12-03 17:38:32.510436961 +0000 UTC m=+1529.327907920" watchObservedRunningTime="2025-12-03 17:38:32.511448849 +0000 UTC m=+1529.328919808" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.519694 4787 scope.go:117] "RemoveContainer" containerID="b76ae34ba03010e44743be5f54b8fd5010504ad6a5669fee127928bb6bd8634c" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.545594 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jlk65"] Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.562375 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7756b9d78c-jlk65"] Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.572051 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.27008415 podStartE2EDuration="11.571996858s" podCreationTimestamp="2025-12-03 17:38:21 +0000 UTC" firstStartedPulling="2025-12-03 17:38:22.750246547 +0000 UTC m=+1519.567717506" lastFinishedPulling="2025-12-03 17:38:32.052159245 +0000 UTC m=+1528.869630214" observedRunningTime="2025-12-03 17:38:32.570943929 +0000 UTC m=+1529.388414888" watchObservedRunningTime="2025-12-03 17:38:32.571996858 +0000 UTC m=+1529.389467817" Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.739536 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.739974 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="85e3cc25-5c94-46d1-a107-a3258de53909" containerName="nova-api-log" containerID="cri-o://38642b74aa386c284e7955a39728d7306137754244f5832b47ecca3313e1b0f5" gracePeriod=30 Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.740107 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="85e3cc25-5c94-46d1-a107-a3258de53909" containerName="nova-api-api" containerID="cri-o://bee944dc494d18fd1543317d49ac4b87f18be6917f77605cd6e8485043777489" gracePeriod=30 Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.770079 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.770330 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="82e8b07e-6b53-4ae7-8570-fa1dc92bd975" containerName="nova-metadata-log" containerID="cri-o://1b3f121048d270d93b69e86c769d4f15d847505815a668ea1cab35e3cfeaa06f" gracePeriod=30 Dec 03 17:38:32 crc kubenswrapper[4787]: I1203 17:38:32.770379 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="82e8b07e-6b53-4ae7-8570-fa1dc92bd975" containerName="nova-metadata-metadata" containerID="cri-o://35fcd12546c3bc57ab7c66269b709ed60bab9892202b0be9bee36a2f23dd4843" gracePeriod=30 Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.512340 4787 generic.go:334] "Generic (PLEG): container finished" podID="85e3cc25-5c94-46d1-a107-a3258de53909" containerID="bee944dc494d18fd1543317d49ac4b87f18be6917f77605cd6e8485043777489" exitCode=0 Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.512613 4787 generic.go:334] "Generic (PLEG): container finished" podID="85e3cc25-5c94-46d1-a107-a3258de53909" containerID="38642b74aa386c284e7955a39728d7306137754244f5832b47ecca3313e1b0f5" exitCode=143 Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.512687 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85e3cc25-5c94-46d1-a107-a3258de53909","Type":"ContainerDied","Data":"bee944dc494d18fd1543317d49ac4b87f18be6917f77605cd6e8485043777489"} Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.512709 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85e3cc25-5c94-46d1-a107-a3258de53909","Type":"ContainerDied","Data":"38642b74aa386c284e7955a39728d7306137754244f5832b47ecca3313e1b0f5"} Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.515569 4787 generic.go:334] "Generic (PLEG): container finished" podID="82e8b07e-6b53-4ae7-8570-fa1dc92bd975" containerID="35fcd12546c3bc57ab7c66269b709ed60bab9892202b0be9bee36a2f23dd4843" exitCode=0 Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.515601 4787 generic.go:334] "Generic (PLEG): container finished" podID="82e8b07e-6b53-4ae7-8570-fa1dc92bd975" containerID="1b3f121048d270d93b69e86c769d4f15d847505815a668ea1cab35e3cfeaa06f" exitCode=143 Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.516148 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82e8b07e-6b53-4ae7-8570-fa1dc92bd975","Type":"ContainerDied","Data":"35fcd12546c3bc57ab7c66269b709ed60bab9892202b0be9bee36a2f23dd4843"} Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.516221 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82e8b07e-6b53-4ae7-8570-fa1dc92bd975","Type":"ContainerDied","Data":"1b3f121048d270d93b69e86c769d4f15d847505815a668ea1cab35e3cfeaa06f"} Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.516638 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.781682 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b67a4816-85d5-404a-a22d-6cfbc455d286" path="/var/lib/kubelet/pods/b67a4816-85d5-404a-a22d-6cfbc455d286/volumes" Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.870681 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.879953 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.990367 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-combined-ca-bundle\") pod \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.990427 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-combined-ca-bundle\") pod \"85e3cc25-5c94-46d1-a107-a3258de53909\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.990484 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmzkg\" (UniqueName: \"kubernetes.io/projected/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-kube-api-access-zmzkg\") pod \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.990574 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85e3cc25-5c94-46d1-a107-a3258de53909-logs\") pod \"85e3cc25-5c94-46d1-a107-a3258de53909\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.990603 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jc4b2\" (UniqueName: \"kubernetes.io/projected/85e3cc25-5c94-46d1-a107-a3258de53909-kube-api-access-jc4b2\") pod \"85e3cc25-5c94-46d1-a107-a3258de53909\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.990754 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-config-data\") pod \"85e3cc25-5c94-46d1-a107-a3258de53909\" (UID: \"85e3cc25-5c94-46d1-a107-a3258de53909\") " Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.990800 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-nova-metadata-tls-certs\") pod \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.990822 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-logs\") pod \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.990866 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-config-data\") pod \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\" (UID: \"82e8b07e-6b53-4ae7-8570-fa1dc92bd975\") " Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.992448 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/85e3cc25-5c94-46d1-a107-a3258de53909-logs" (OuterVolumeSpecName: "logs") pod "85e3cc25-5c94-46d1-a107-a3258de53909" (UID: "85e3cc25-5c94-46d1-a107-a3258de53909"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.992570 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-logs" (OuterVolumeSpecName: "logs") pod "82e8b07e-6b53-4ae7-8570-fa1dc92bd975" (UID: "82e8b07e-6b53-4ae7-8570-fa1dc92bd975"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:33 crc kubenswrapper[4787]: I1203 17:38:33.999114 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-kube-api-access-zmzkg" (OuterVolumeSpecName: "kube-api-access-zmzkg") pod "82e8b07e-6b53-4ae7-8570-fa1dc92bd975" (UID: "82e8b07e-6b53-4ae7-8570-fa1dc92bd975"). InnerVolumeSpecName "kube-api-access-zmzkg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.025140 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/85e3cc25-5c94-46d1-a107-a3258de53909-kube-api-access-jc4b2" (OuterVolumeSpecName: "kube-api-access-jc4b2") pod "85e3cc25-5c94-46d1-a107-a3258de53909" (UID: "85e3cc25-5c94-46d1-a107-a3258de53909"). InnerVolumeSpecName "kube-api-access-jc4b2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.031804 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-config-data" (OuterVolumeSpecName: "config-data") pod "82e8b07e-6b53-4ae7-8570-fa1dc92bd975" (UID: "82e8b07e-6b53-4ae7-8570-fa1dc92bd975"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.036269 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-config-data" (OuterVolumeSpecName: "config-data") pod "85e3cc25-5c94-46d1-a107-a3258de53909" (UID: "85e3cc25-5c94-46d1-a107-a3258de53909"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.054417 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "85e3cc25-5c94-46d1-a107-a3258de53909" (UID: "85e3cc25-5c94-46d1-a107-a3258de53909"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.061380 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "82e8b07e-6b53-4ae7-8570-fa1dc92bd975" (UID: "82e8b07e-6b53-4ae7-8570-fa1dc92bd975"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.063979 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "82e8b07e-6b53-4ae7-8570-fa1dc92bd975" (UID: "82e8b07e-6b53-4ae7-8570-fa1dc92bd975"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.097253 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.097299 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.097312 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmzkg\" (UniqueName: \"kubernetes.io/projected/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-kube-api-access-zmzkg\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.097326 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/85e3cc25-5c94-46d1-a107-a3258de53909-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.097338 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jc4b2\" (UniqueName: \"kubernetes.io/projected/85e3cc25-5c94-46d1-a107-a3258de53909-kube-api-access-jc4b2\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.097350 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/85e3cc25-5c94-46d1-a107-a3258de53909-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.097364 4787 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.097377 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.097390 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/82e8b07e-6b53-4ae7-8570-fa1dc92bd975-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.528280 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.528281 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"82e8b07e-6b53-4ae7-8570-fa1dc92bd975","Type":"ContainerDied","Data":"4691eb61c2989b562475460aedde64faf9f0efa1058ca5d3a29c3f9179e079b7"} Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.528417 4787 scope.go:117] "RemoveContainer" containerID="35fcd12546c3bc57ab7c66269b709ed60bab9892202b0be9bee36a2f23dd4843" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.535229 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"85e3cc25-5c94-46d1-a107-a3258de53909","Type":"ContainerDied","Data":"4943edae8da058bde0b89216c43952f0bb34251e85735bff2ce9ef31e0bee853"} Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.536803 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.561407 4787 scope.go:117] "RemoveContainer" containerID="1b3f121048d270d93b69e86c769d4f15d847505815a668ea1cab35e3cfeaa06f" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.594753 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.618555 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.620931 4787 scope.go:117] "RemoveContainer" containerID="bee944dc494d18fd1543317d49ac4b87f18be6917f77605cd6e8485043777489" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.656069 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.668426 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.676298 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:34 crc kubenswrapper[4787]: E1203 17:38:34.676853 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82e8b07e-6b53-4ae7-8570-fa1dc92bd975" containerName="nova-metadata-metadata" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.676873 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="82e8b07e-6b53-4ae7-8570-fa1dc92bd975" containerName="nova-metadata-metadata" Dec 03 17:38:34 crc kubenswrapper[4787]: E1203 17:38:34.676891 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a5f00a92-19b0-4f7b-a3b8-66b3e720fab3" containerName="nova-manage" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.676899 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a5f00a92-19b0-4f7b-a3b8-66b3e720fab3" containerName="nova-manage" Dec 03 17:38:34 crc kubenswrapper[4787]: E1203 17:38:34.676933 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b67a4816-85d5-404a-a22d-6cfbc455d286" containerName="dnsmasq-dns" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.676940 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b67a4816-85d5-404a-a22d-6cfbc455d286" containerName="dnsmasq-dns" Dec 03 17:38:34 crc kubenswrapper[4787]: E1203 17:38:34.676956 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b67a4816-85d5-404a-a22d-6cfbc455d286" containerName="init" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.676963 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b67a4816-85d5-404a-a22d-6cfbc455d286" containerName="init" Dec 03 17:38:34 crc kubenswrapper[4787]: E1203 17:38:34.676973 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85e3cc25-5c94-46d1-a107-a3258de53909" containerName="nova-api-api" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.676980 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="85e3cc25-5c94-46d1-a107-a3258de53909" containerName="nova-api-api" Dec 03 17:38:34 crc kubenswrapper[4787]: E1203 17:38:34.676993 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82e8b07e-6b53-4ae7-8570-fa1dc92bd975" containerName="nova-metadata-log" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.677000 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="82e8b07e-6b53-4ae7-8570-fa1dc92bd975" containerName="nova-metadata-log" Dec 03 17:38:34 crc kubenswrapper[4787]: E1203 17:38:34.677011 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="85e3cc25-5c94-46d1-a107-a3258de53909" containerName="nova-api-log" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.677036 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="85e3cc25-5c94-46d1-a107-a3258de53909" containerName="nova-api-log" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.677283 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="85e3cc25-5c94-46d1-a107-a3258de53909" containerName="nova-api-log" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.677307 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a5f00a92-19b0-4f7b-a3b8-66b3e720fab3" containerName="nova-manage" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.677333 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="82e8b07e-6b53-4ae7-8570-fa1dc92bd975" containerName="nova-metadata-log" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.677350 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="82e8b07e-6b53-4ae7-8570-fa1dc92bd975" containerName="nova-metadata-metadata" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.677365 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b67a4816-85d5-404a-a22d-6cfbc455d286" containerName="dnsmasq-dns" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.677380 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="85e3cc25-5c94-46d1-a107-a3258de53909" containerName="nova-api-api" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.678977 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.681730 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.682159 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.697355 4787 scope.go:117] "RemoveContainer" containerID="38642b74aa386c284e7955a39728d7306137754244f5832b47ecca3313e1b0f5" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.697446 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.699856 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.702886 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.713278 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.726643 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.820442 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.820761 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-config-data\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.820917 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzb9v\" (UniqueName: \"kubernetes.io/projected/695faa7a-8e43-4277-b206-4c27ae5fd3a5-kube-api-access-mzb9v\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.820993 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shd4h\" (UniqueName: \"kubernetes.io/projected/2d4b067d-cfb3-4923-ae34-0cdee637eb81-kube-api-access-shd4h\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.821149 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.821191 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.821218 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/695faa7a-8e43-4277-b206-4c27ae5fd3a5-logs\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.821388 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d4b067d-cfb3-4923-ae34-0cdee637eb81-logs\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.821433 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-config-data\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.923821 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shd4h\" (UniqueName: \"kubernetes.io/projected/2d4b067d-cfb3-4923-ae34-0cdee637eb81-kube-api-access-shd4h\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.923966 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.924061 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.924100 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/695faa7a-8e43-4277-b206-4c27ae5fd3a5-logs\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.924195 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d4b067d-cfb3-4923-ae34-0cdee637eb81-logs\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.924230 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-config-data\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.924352 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.924418 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-config-data\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.924558 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzb9v\" (UniqueName: \"kubernetes.io/projected/695faa7a-8e43-4277-b206-4c27ae5fd3a5-kube-api-access-mzb9v\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.924833 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d4b067d-cfb3-4923-ae34-0cdee637eb81-logs\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.924960 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/695faa7a-8e43-4277-b206-4c27ae5fd3a5-logs\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.928778 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.931183 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.934669 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.935442 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-config-data\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.939048 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shd4h\" (UniqueName: \"kubernetes.io/projected/2d4b067d-cfb3-4923-ae34-0cdee637eb81-kube-api-access-shd4h\") pod \"nova-api-0\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " pod="openstack/nova-api-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.940125 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-config-data\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:34 crc kubenswrapper[4787]: I1203 17:38:34.942834 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzb9v\" (UniqueName: \"kubernetes.io/projected/695faa7a-8e43-4277-b206-4c27ae5fd3a5-kube-api-access-mzb9v\") pod \"nova-metadata-0\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " pod="openstack/nova-metadata-0" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.004948 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.082115 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.378240 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pwxs4"] Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.383986 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.411295 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pwxs4"] Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.511120 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:38:35 crc kubenswrapper[4787]: W1203 17:38:35.515563 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod695faa7a_8e43_4277_b206_4c27ae5fd3a5.slice/crio-a27829b13ab47b4cedd6e0e3cadfe60456e74eb38a13383d361d6c6aee49dcf6 WatchSource:0}: Error finding container a27829b13ab47b4cedd6e0e3cadfe60456e74eb38a13383d361d6c6aee49dcf6: Status 404 returned error can't find the container with id a27829b13ab47b4cedd6e0e3cadfe60456e74eb38a13383d361d6c6aee49dcf6 Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.538003 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-catalog-content\") pod \"redhat-marketplace-pwxs4\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.538093 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-utilities\") pod \"redhat-marketplace-pwxs4\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.538127 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jftpb\" (UniqueName: \"kubernetes.io/projected/8d10b8a6-f576-452e-8fe1-14908ea925b3-kube-api-access-jftpb\") pod \"redhat-marketplace-pwxs4\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.545563 4787 generic.go:334] "Generic (PLEG): container finished" podID="e43d4111-655b-4843-bdc1-1465d8c5a050" containerID="044d86c2a2fe3dbd8a3c6756e51f420f54addb310c01feaa7100fba82c302de5" exitCode=0 Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.545649 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-45bq2" event={"ID":"e43d4111-655b-4843-bdc1-1465d8c5a050","Type":"ContainerDied","Data":"044d86c2a2fe3dbd8a3c6756e51f420f54addb310c01feaa7100fba82c302de5"} Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.549686 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"695faa7a-8e43-4277-b206-4c27ae5fd3a5","Type":"ContainerStarted","Data":"a27829b13ab47b4cedd6e0e3cadfe60456e74eb38a13383d361d6c6aee49dcf6"} Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.554995 4787 generic.go:334] "Generic (PLEG): container finished" podID="92bdbf6e-87dc-4a93-aefd-26b002566663" containerID="021fa014eff70f032132ca3763ca662441515e52086cce85320c771165ecb316" exitCode=0 Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.555201 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-8jchz" event={"ID":"92bdbf6e-87dc-4a93-aefd-26b002566663","Type":"ContainerDied","Data":"021fa014eff70f032132ca3763ca662441515e52086cce85320c771165ecb316"} Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.640870 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-catalog-content\") pod \"redhat-marketplace-pwxs4\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.640931 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-utilities\") pod \"redhat-marketplace-pwxs4\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.640957 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jftpb\" (UniqueName: \"kubernetes.io/projected/8d10b8a6-f576-452e-8fe1-14908ea925b3-kube-api-access-jftpb\") pod \"redhat-marketplace-pwxs4\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.641381 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-catalog-content\") pod \"redhat-marketplace-pwxs4\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.641469 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-utilities\") pod \"redhat-marketplace-pwxs4\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.670009 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jftpb\" (UniqueName: \"kubernetes.io/projected/8d10b8a6-f576-452e-8fe1-14908ea925b3-kube-api-access-jftpb\") pod \"redhat-marketplace-pwxs4\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.670520 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.718307 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.784314 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82e8b07e-6b53-4ae7-8570-fa1dc92bd975" path="/var/lib/kubelet/pods/82e8b07e-6b53-4ae7-8570-fa1dc92bd975/volumes" Dec 03 17:38:35 crc kubenswrapper[4787]: I1203 17:38:35.785096 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="85e3cc25-5c94-46d1-a107-a3258de53909" path="/var/lib/kubelet/pods/85e3cc25-5c94-46d1-a107-a3258de53909/volumes" Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.222319 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pwxs4"] Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.582740 4787 generic.go:334] "Generic (PLEG): container finished" podID="8d10b8a6-f576-452e-8fe1-14908ea925b3" containerID="5576f8717aaeb4a23c47314f1af15835aa4f498a16dd1293132c8193609f8127" exitCode=0 Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.583197 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwxs4" event={"ID":"8d10b8a6-f576-452e-8fe1-14908ea925b3","Type":"ContainerDied","Data":"5576f8717aaeb4a23c47314f1af15835aa4f498a16dd1293132c8193609f8127"} Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.583243 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwxs4" event={"ID":"8d10b8a6-f576-452e-8fe1-14908ea925b3","Type":"ContainerStarted","Data":"0bc9a6f479f79dd9065e6ceb81a9cd91ae2fe055c7e635bdf578134b09742727"} Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.587680 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d4b067d-cfb3-4923-ae34-0cdee637eb81","Type":"ContainerStarted","Data":"df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d"} Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.587928 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d4b067d-cfb3-4923-ae34-0cdee637eb81","Type":"ContainerStarted","Data":"8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499"} Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.588326 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d4b067d-cfb3-4923-ae34-0cdee637eb81","Type":"ContainerStarted","Data":"e2f1198476ba2af04e100ad194b3485679b9e7e8d999da6216107837905a9563"} Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.603332 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"695faa7a-8e43-4277-b206-4c27ae5fd3a5","Type":"ContainerStarted","Data":"8d7b157b2eb22560a8597b21b7b03924b206ff3e05a9c5677d380b3234867a49"} Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.603410 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"695faa7a-8e43-4277-b206-4c27ae5fd3a5","Type":"ContainerStarted","Data":"8b1452abe4a84ea337f4eca5edd118c0aaab837bc5b6d8ae4fe654f7a14ef2b4"} Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.649298 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.649274552 podStartE2EDuration="2.649274552s" podCreationTimestamp="2025-12-03 17:38:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:36.640782742 +0000 UTC m=+1533.458253701" watchObservedRunningTime="2025-12-03 17:38:36.649274552 +0000 UTC m=+1533.466745521" Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.674152 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.674110634 podStartE2EDuration="2.674110634s" podCreationTimestamp="2025-12-03 17:38:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:36.65993072 +0000 UTC m=+1533.477401679" watchObservedRunningTime="2025-12-03 17:38:36.674110634 +0000 UTC m=+1533.491581593" Dec 03 17:38:36 crc kubenswrapper[4787]: I1203 17:38:36.819466 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-7756b9d78c-jlk65" podUID="b67a4816-85d5-404a-a22d-6cfbc455d286" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.192:5353: i/o timeout" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.182039 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.194540 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.288239 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-config-data\") pod \"92bdbf6e-87dc-4a93-aefd-26b002566663\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.288290 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-scripts\") pod \"92bdbf6e-87dc-4a93-aefd-26b002566663\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.288409 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-scripts\") pod \"e43d4111-655b-4843-bdc1-1465d8c5a050\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.288432 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-combined-ca-bundle\") pod \"92bdbf6e-87dc-4a93-aefd-26b002566663\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.288464 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-combined-ca-bundle\") pod \"e43d4111-655b-4843-bdc1-1465d8c5a050\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.288590 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npxt9\" (UniqueName: \"kubernetes.io/projected/92bdbf6e-87dc-4a93-aefd-26b002566663-kube-api-access-npxt9\") pod \"92bdbf6e-87dc-4a93-aefd-26b002566663\" (UID: \"92bdbf6e-87dc-4a93-aefd-26b002566663\") " Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.288629 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-config-data\") pod \"e43d4111-655b-4843-bdc1-1465d8c5a050\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.288657 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9kbvg\" (UniqueName: \"kubernetes.io/projected/e43d4111-655b-4843-bdc1-1465d8c5a050-kube-api-access-9kbvg\") pod \"e43d4111-655b-4843-bdc1-1465d8c5a050\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.294391 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92bdbf6e-87dc-4a93-aefd-26b002566663-kube-api-access-npxt9" (OuterVolumeSpecName: "kube-api-access-npxt9") pod "92bdbf6e-87dc-4a93-aefd-26b002566663" (UID: "92bdbf6e-87dc-4a93-aefd-26b002566663"). InnerVolumeSpecName "kube-api-access-npxt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.294511 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-scripts" (OuterVolumeSpecName: "scripts") pod "e43d4111-655b-4843-bdc1-1465d8c5a050" (UID: "e43d4111-655b-4843-bdc1-1465d8c5a050"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.299297 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e43d4111-655b-4843-bdc1-1465d8c5a050-kube-api-access-9kbvg" (OuterVolumeSpecName: "kube-api-access-9kbvg") pod "e43d4111-655b-4843-bdc1-1465d8c5a050" (UID: "e43d4111-655b-4843-bdc1-1465d8c5a050"). InnerVolumeSpecName "kube-api-access-9kbvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.317156 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-scripts" (OuterVolumeSpecName: "scripts") pod "92bdbf6e-87dc-4a93-aefd-26b002566663" (UID: "92bdbf6e-87dc-4a93-aefd-26b002566663"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.318177 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-config-data" (OuterVolumeSpecName: "config-data") pod "92bdbf6e-87dc-4a93-aefd-26b002566663" (UID: "92bdbf6e-87dc-4a93-aefd-26b002566663"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:37 crc kubenswrapper[4787]: E1203 17:38:37.321679 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-combined-ca-bundle podName:e43d4111-655b-4843-bdc1-1465d8c5a050 nodeName:}" failed. No retries permitted until 2025-12-03 17:38:37.821649333 +0000 UTC m=+1534.639120282 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-combined-ca-bundle") pod "e43d4111-655b-4843-bdc1-1465d8c5a050" (UID: "e43d4111-655b-4843-bdc1-1465d8c5a050") : error deleting /var/lib/kubelet/pods/e43d4111-655b-4843-bdc1-1465d8c5a050/volume-subpaths: remove /var/lib/kubelet/pods/e43d4111-655b-4843-bdc1-1465d8c5a050/volume-subpaths: no such file or directory Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.324506 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-config-data" (OuterVolumeSpecName: "config-data") pod "e43d4111-655b-4843-bdc1-1465d8c5a050" (UID: "e43d4111-655b-4843-bdc1-1465d8c5a050"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.325649 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92bdbf6e-87dc-4a93-aefd-26b002566663" (UID: "92bdbf6e-87dc-4a93-aefd-26b002566663"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.393643 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.393686 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9kbvg\" (UniqueName: \"kubernetes.io/projected/e43d4111-655b-4843-bdc1-1465d8c5a050-kube-api-access-9kbvg\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.393701 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.393713 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.393724 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.393736 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92bdbf6e-87dc-4a93-aefd-26b002566663-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.393749 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npxt9\" (UniqueName: \"kubernetes.io/projected/92bdbf6e-87dc-4a93-aefd-26b002566663-kube-api-access-npxt9\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.622142 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-8jchz" event={"ID":"92bdbf6e-87dc-4a93-aefd-26b002566663","Type":"ContainerDied","Data":"16ec847b2ca36ce1524408f49589720b3d3f4827e7e94138533ceb4c2a4dc841"} Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.622503 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16ec847b2ca36ce1524408f49589720b3d3f4827e7e94138533ceb4c2a4dc841" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.622815 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-8jchz" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.629749 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwxs4" event={"ID":"8d10b8a6-f576-452e-8fe1-14908ea925b3","Type":"ContainerStarted","Data":"57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a"} Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.650302 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-45bq2" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.650875 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-45bq2" event={"ID":"e43d4111-655b-4843-bdc1-1465d8c5a050","Type":"ContainerDied","Data":"f1b50ca9a7e0b1da415f88e6e67d20e9ccfadc216f74b37be46da321b23d8e88"} Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.650910 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f1b50ca9a7e0b1da415f88e6e67d20e9ccfadc216f74b37be46da321b23d8e88" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.678253 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 17:38:37 crc kubenswrapper[4787]: E1203 17:38:37.678862 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e43d4111-655b-4843-bdc1-1465d8c5a050" containerName="nova-cell1-conductor-db-sync" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.678890 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e43d4111-655b-4843-bdc1-1465d8c5a050" containerName="nova-cell1-conductor-db-sync" Dec 03 17:38:37 crc kubenswrapper[4787]: E1203 17:38:37.678914 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92bdbf6e-87dc-4a93-aefd-26b002566663" containerName="aodh-db-sync" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.678923 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="92bdbf6e-87dc-4a93-aefd-26b002566663" containerName="aodh-db-sync" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.679183 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e43d4111-655b-4843-bdc1-1465d8c5a050" containerName="nova-cell1-conductor-db-sync" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.679222 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="92bdbf6e-87dc-4a93-aefd-26b002566663" containerName="aodh-db-sync" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.680221 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.697764 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.805132 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d63cec-994c-4594-8521-c9e8b5252363-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"70d63cec-994c-4594-8521-c9e8b5252363\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.805179 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d63cec-994c-4594-8521-c9e8b5252363-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"70d63cec-994c-4594-8521-c9e8b5252363\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.805218 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngwxr\" (UniqueName: \"kubernetes.io/projected/70d63cec-994c-4594-8521-c9e8b5252363-kube-api-access-ngwxr\") pod \"nova-cell1-conductor-0\" (UID: \"70d63cec-994c-4594-8521-c9e8b5252363\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.908541 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-combined-ca-bundle\") pod \"e43d4111-655b-4843-bdc1-1465d8c5a050\" (UID: \"e43d4111-655b-4843-bdc1-1465d8c5a050\") " Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.910859 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d63cec-994c-4594-8521-c9e8b5252363-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"70d63cec-994c-4594-8521-c9e8b5252363\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.910916 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d63cec-994c-4594-8521-c9e8b5252363-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"70d63cec-994c-4594-8521-c9e8b5252363\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.912405 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngwxr\" (UniqueName: \"kubernetes.io/projected/70d63cec-994c-4594-8521-c9e8b5252363-kube-api-access-ngwxr\") pod \"nova-cell1-conductor-0\" (UID: \"70d63cec-994c-4594-8521-c9e8b5252363\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.916204 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/70d63cec-994c-4594-8521-c9e8b5252363-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"70d63cec-994c-4594-8521-c9e8b5252363\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.920236 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e43d4111-655b-4843-bdc1-1465d8c5a050" (UID: "e43d4111-655b-4843-bdc1-1465d8c5a050"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.925108 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/70d63cec-994c-4594-8521-c9e8b5252363-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"70d63cec-994c-4594-8521-c9e8b5252363\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:37 crc kubenswrapper[4787]: I1203 17:38:37.938398 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngwxr\" (UniqueName: \"kubernetes.io/projected/70d63cec-994c-4594-8521-c9e8b5252363-kube-api-access-ngwxr\") pod \"nova-cell1-conductor-0\" (UID: \"70d63cec-994c-4594-8521-c9e8b5252363\") " pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.001201 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.018499 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e43d4111-655b-4843-bdc1-1465d8c5a050-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.040855 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.043573 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.061615 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.063237 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.064792 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-q4n2s" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.099783 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.121380 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4n7j8\" (UniqueName: \"kubernetes.io/projected/4282e0c2-7846-4f1f-afeb-076726cfe29d-kube-api-access-4n7j8\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.121727 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.121822 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-scripts\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.121929 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-config-data\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.223642 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-scripts\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.223817 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-config-data\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.223873 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4n7j8\" (UniqueName: \"kubernetes.io/projected/4282e0c2-7846-4f1f-afeb-076726cfe29d-kube-api-access-4n7j8\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.223900 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.234061 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-combined-ca-bundle\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.238180 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-config-data\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.245706 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-scripts\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.265506 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4n7j8\" (UniqueName: \"kubernetes.io/projected/4282e0c2-7846-4f1f-afeb-076726cfe29d-kube-api-access-4n7j8\") pod \"aodh-0\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.468999 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.605931 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 03 17:38:38 crc kubenswrapper[4787]: W1203 17:38:38.608987 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70d63cec_994c_4594_8521_c9e8b5252363.slice/crio-4ce2bd1049d5eb549696a57deda044f2dc470c93c4ea21ab772e3bba18bae2e4 WatchSource:0}: Error finding container 4ce2bd1049d5eb549696a57deda044f2dc470c93c4ea21ab772e3bba18bae2e4: Status 404 returned error can't find the container with id 4ce2bd1049d5eb549696a57deda044f2dc470c93c4ea21ab772e3bba18bae2e4 Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.682743 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"70d63cec-994c-4594-8521-c9e8b5252363","Type":"ContainerStarted","Data":"4ce2bd1049d5eb549696a57deda044f2dc470c93c4ea21ab772e3bba18bae2e4"} Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.689968 4787 generic.go:334] "Generic (PLEG): container finished" podID="8d10b8a6-f576-452e-8fe1-14908ea925b3" containerID="57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a" exitCode=0 Dec 03 17:38:38 crc kubenswrapper[4787]: I1203 17:38:38.690010 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwxs4" event={"ID":"8d10b8a6-f576-452e-8fe1-14908ea925b3","Type":"ContainerDied","Data":"57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a"} Dec 03 17:38:39 crc kubenswrapper[4787]: W1203 17:38:39.010478 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4282e0c2_7846_4f1f_afeb_076726cfe29d.slice/crio-a6093dcf7ba99c69f3545111f2d8b1f0472a6d621770fb905f7a7891af96b1ff WatchSource:0}: Error finding container a6093dcf7ba99c69f3545111f2d8b1f0472a6d621770fb905f7a7891af96b1ff: Status 404 returned error can't find the container with id a6093dcf7ba99c69f3545111f2d8b1f0472a6d621770fb905f7a7891af96b1ff Dec 03 17:38:39 crc kubenswrapper[4787]: I1203 17:38:39.012478 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 17:38:39 crc kubenswrapper[4787]: I1203 17:38:39.701780 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4282e0c2-7846-4f1f-afeb-076726cfe29d","Type":"ContainerStarted","Data":"a6093dcf7ba99c69f3545111f2d8b1f0472a6d621770fb905f7a7891af96b1ff"} Dec 03 17:38:40 crc kubenswrapper[4787]: I1203 17:38:40.005736 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:38:40 crc kubenswrapper[4787]: I1203 17:38:40.006116 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:38:40 crc kubenswrapper[4787]: I1203 17:38:40.714780 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"70d63cec-994c-4594-8521-c9e8b5252363","Type":"ContainerStarted","Data":"1580cba514fa13d134e59e71f2d78daa42f0010dfa7258fd1c70a9eb2b5b3dab"} Dec 03 17:38:40 crc kubenswrapper[4787]: I1203 17:38:40.738190 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.738169221 podStartE2EDuration="3.738169221s" podCreationTimestamp="2025-12-03 17:38:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:40.731220243 +0000 UTC m=+1537.548691212" watchObservedRunningTime="2025-12-03 17:38:40.738169221 +0000 UTC m=+1537.555640180" Dec 03 17:38:41 crc kubenswrapper[4787]: I1203 17:38:41.036478 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 17:38:41 crc kubenswrapper[4787]: I1203 17:38:41.745592 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4282e0c2-7846-4f1f-afeb-076726cfe29d","Type":"ContainerStarted","Data":"fc7909877fcab48f36abd4f3e735bc3e386dede90fc29fb0260a7deda5ac65b4"} Dec 03 17:38:41 crc kubenswrapper[4787]: I1203 17:38:41.762381 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwxs4" event={"ID":"8d10b8a6-f576-452e-8fe1-14908ea925b3","Type":"ContainerStarted","Data":"18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03"} Dec 03 17:38:41 crc kubenswrapper[4787]: I1203 17:38:41.762459 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:41 crc kubenswrapper[4787]: I1203 17:38:41.830165 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pwxs4" podStartSLOduration=2.831301579 podStartE2EDuration="6.83013965s" podCreationTimestamp="2025-12-03 17:38:35 +0000 UTC" firstStartedPulling="2025-12-03 17:38:36.587785267 +0000 UTC m=+1533.405256226" lastFinishedPulling="2025-12-03 17:38:40.586623338 +0000 UTC m=+1537.404094297" observedRunningTime="2025-12-03 17:38:41.814069885 +0000 UTC m=+1538.631540844" watchObservedRunningTime="2025-12-03 17:38:41.83013965 +0000 UTC m=+1538.647610609" Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.207734 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.208452 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="proxy-httpd" containerID="cri-o://715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65" gracePeriod=30 Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.208674 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="sg-core" containerID="cri-o://d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d" gracePeriod=30 Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.208745 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="ceilometer-notification-agent" containerID="cri-o://35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e" gracePeriod=30 Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.208374 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="ceilometer-central-agent" containerID="cri-o://73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420" gracePeriod=30 Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.216192 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.793174 4787 generic.go:334] "Generic (PLEG): container finished" podID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerID="715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65" exitCode=0 Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.793546 4787 generic.go:334] "Generic (PLEG): container finished" podID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerID="d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d" exitCode=2 Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.793558 4787 generic.go:334] "Generic (PLEG): container finished" podID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerID="73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420" exitCode=0 Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.793601 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"758c0073-2001-4fa7-b1f1-e5350aae3e33","Type":"ContainerDied","Data":"715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65"} Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.793654 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"758c0073-2001-4fa7-b1f1-e5350aae3e33","Type":"ContainerDied","Data":"d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d"} Dec 03 17:38:43 crc kubenswrapper[4787]: I1203 17:38:43.793728 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"758c0073-2001-4fa7-b1f1-e5350aae3e33","Type":"ContainerDied","Data":"73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420"} Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.254566 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.392485 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-combined-ca-bundle\") pod \"758c0073-2001-4fa7-b1f1-e5350aae3e33\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.392592 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-log-httpd\") pod \"758c0073-2001-4fa7-b1f1-e5350aae3e33\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.392728 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5kl8\" (UniqueName: \"kubernetes.io/projected/758c0073-2001-4fa7-b1f1-e5350aae3e33-kube-api-access-p5kl8\") pod \"758c0073-2001-4fa7-b1f1-e5350aae3e33\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.392770 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-scripts\") pod \"758c0073-2001-4fa7-b1f1-e5350aae3e33\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.392829 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-config-data\") pod \"758c0073-2001-4fa7-b1f1-e5350aae3e33\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.392983 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-sg-core-conf-yaml\") pod \"758c0073-2001-4fa7-b1f1-e5350aae3e33\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.393061 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-run-httpd\") pod \"758c0073-2001-4fa7-b1f1-e5350aae3e33\" (UID: \"758c0073-2001-4fa7-b1f1-e5350aae3e33\") " Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.393331 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "758c0073-2001-4fa7-b1f1-e5350aae3e33" (UID: "758c0073-2001-4fa7-b1f1-e5350aae3e33"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.393822 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "758c0073-2001-4fa7-b1f1-e5350aae3e33" (UID: "758c0073-2001-4fa7-b1f1-e5350aae3e33"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.394160 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.394689 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/758c0073-2001-4fa7-b1f1-e5350aae3e33-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.420790 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-scripts" (OuterVolumeSpecName: "scripts") pod "758c0073-2001-4fa7-b1f1-e5350aae3e33" (UID: "758c0073-2001-4fa7-b1f1-e5350aae3e33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.426259 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/758c0073-2001-4fa7-b1f1-e5350aae3e33-kube-api-access-p5kl8" (OuterVolumeSpecName: "kube-api-access-p5kl8") pod "758c0073-2001-4fa7-b1f1-e5350aae3e33" (UID: "758c0073-2001-4fa7-b1f1-e5350aae3e33"). InnerVolumeSpecName "kube-api-access-p5kl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.500415 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5kl8\" (UniqueName: \"kubernetes.io/projected/758c0073-2001-4fa7-b1f1-e5350aae3e33-kube-api-access-p5kl8\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.500468 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.507178 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "758c0073-2001-4fa7-b1f1-e5350aae3e33" (UID: "758c0073-2001-4fa7-b1f1-e5350aae3e33"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.599835 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "758c0073-2001-4fa7-b1f1-e5350aae3e33" (UID: "758c0073-2001-4fa7-b1f1-e5350aae3e33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.602245 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.602273 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.622079 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-config-data" (OuterVolumeSpecName: "config-data") pod "758c0073-2001-4fa7-b1f1-e5350aae3e33" (UID: "758c0073-2001-4fa7-b1f1-e5350aae3e33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.703938 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/758c0073-2001-4fa7-b1f1-e5350aae3e33-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.804970 4787 generic.go:334] "Generic (PLEG): container finished" podID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerID="35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e" exitCode=0 Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.805052 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"758c0073-2001-4fa7-b1f1-e5350aae3e33","Type":"ContainerDied","Data":"35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e"} Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.805443 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"758c0073-2001-4fa7-b1f1-e5350aae3e33","Type":"ContainerDied","Data":"9f68b33ba2baf7d6659a5799041e1aa5772c9aaff312161bbff190bf4bbd784f"} Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.805068 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.805494 4787 scope.go:117] "RemoveContainer" containerID="715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.807457 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4282e0c2-7846-4f1f-afeb-076726cfe29d","Type":"ContainerStarted","Data":"1adb2a520fce01177ef81401295860b153f5aab544859ed528f5654fa3326feb"} Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.860214 4787 scope.go:117] "RemoveContainer" containerID="d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.860409 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.875080 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.892086 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:44 crc kubenswrapper[4787]: E1203 17:38:44.892658 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="ceilometer-central-agent" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.892680 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="ceilometer-central-agent" Dec 03 17:38:44 crc kubenswrapper[4787]: E1203 17:38:44.892727 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="proxy-httpd" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.892734 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="proxy-httpd" Dec 03 17:38:44 crc kubenswrapper[4787]: E1203 17:38:44.892747 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="ceilometer-notification-agent" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.892756 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="ceilometer-notification-agent" Dec 03 17:38:44 crc kubenswrapper[4787]: E1203 17:38:44.892779 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="sg-core" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.892787 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="sg-core" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.893051 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="ceilometer-notification-agent" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.893068 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="proxy-httpd" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.893092 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="sg-core" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.893106 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" containerName="ceilometer-central-agent" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.895047 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.899509 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.899742 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.909598 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.912965 4787 scope.go:117] "RemoveContainer" containerID="35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.952826 4787 scope.go:117] "RemoveContainer" containerID="73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.983503 4787 scope.go:117] "RemoveContainer" containerID="715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65" Dec 03 17:38:44 crc kubenswrapper[4787]: E1203 17:38:44.983989 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65\": container with ID starting with 715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65 not found: ID does not exist" containerID="715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.984048 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65"} err="failed to get container status \"715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65\": rpc error: code = NotFound desc = could not find container \"715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65\": container with ID starting with 715e3cf322d47d9b290ef2fe7782a3a844f9719ecae014ae76aa2ae7f6b83a65 not found: ID does not exist" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.984075 4787 scope.go:117] "RemoveContainer" containerID="d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d" Dec 03 17:38:44 crc kubenswrapper[4787]: E1203 17:38:44.984533 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d\": container with ID starting with d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d not found: ID does not exist" containerID="d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.984688 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d"} err="failed to get container status \"d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d\": rpc error: code = NotFound desc = could not find container \"d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d\": container with ID starting with d2bad9779454a128669848a8f05c956931fbe4efb241cd5c9d5fd35ba08d321d not found: ID does not exist" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.984717 4787 scope.go:117] "RemoveContainer" containerID="35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e" Dec 03 17:38:44 crc kubenswrapper[4787]: E1203 17:38:44.985054 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e\": container with ID starting with 35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e not found: ID does not exist" containerID="35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.985082 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e"} err="failed to get container status \"35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e\": rpc error: code = NotFound desc = could not find container \"35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e\": container with ID starting with 35612556dfcdda0d0a18ce6ef9a4b4697f05e21ece2553037d04f5177cd0cf5e not found: ID does not exist" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.985105 4787 scope.go:117] "RemoveContainer" containerID="73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420" Dec 03 17:38:44 crc kubenswrapper[4787]: E1203 17:38:44.985820 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420\": container with ID starting with 73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420 not found: ID does not exist" containerID="73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420" Dec 03 17:38:44 crc kubenswrapper[4787]: I1203 17:38:44.985868 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420"} err="failed to get container status \"73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420\": rpc error: code = NotFound desc = could not find container \"73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420\": container with ID starting with 73b94459a309483f0afe77023e2ab0427ce9b45f27070a7b793e5d504e46a420 not found: ID does not exist" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.006153 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.006233 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.009764 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zkzjv\" (UniqueName: \"kubernetes.io/projected/8331dfaa-fd3d-4609-a078-f16f781e0318-kube-api-access-zkzjv\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.010118 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-config-data\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.010145 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-run-httpd\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.010163 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-scripts\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.010190 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-log-httpd\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.010447 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.010602 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.083709 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.083774 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.112444 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.112538 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.112598 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zkzjv\" (UniqueName: \"kubernetes.io/projected/8331dfaa-fd3d-4609-a078-f16f781e0318-kube-api-access-zkzjv\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.112667 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-config-data\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.112713 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-run-httpd\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.112732 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-scripts\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.112757 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-log-httpd\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.113185 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-log-httpd\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.113257 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-run-httpd\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.117045 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-scripts\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.117098 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.119073 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-config-data\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.131439 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.133738 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zkzjv\" (UniqueName: \"kubernetes.io/projected/8331dfaa-fd3d-4609-a078-f16f781e0318-kube-api-access-zkzjv\") pod \"ceilometer-0\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.218553 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.727891 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.728276 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.787359 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="758c0073-2001-4fa7-b1f1-e5350aae3e33" path="/var/lib/kubelet/pods/758c0073-2001-4fa7-b1f1-e5350aae3e33/volumes" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.789991 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.818047 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:45 crc kubenswrapper[4787]: I1203 17:38:45.879487 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:46 crc kubenswrapper[4787]: I1203 17:38:46.022236 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.230:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:38:46 crc kubenswrapper[4787]: I1203 17:38:46.022593 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.230:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:38:46 crc kubenswrapper[4787]: I1203 17:38:46.034564 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pwxs4"] Dec 03 17:38:46 crc kubenswrapper[4787]: I1203 17:38:46.166217 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.231:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 17:38:46 crc kubenswrapper[4787]: I1203 17:38:46.166533 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.231:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 17:38:46 crc kubenswrapper[4787]: I1203 17:38:46.434942 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:46 crc kubenswrapper[4787]: I1203 17:38:46.844652 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8331dfaa-fd3d-4609-a078-f16f781e0318","Type":"ContainerStarted","Data":"9f6f2df677b9f31d2bb45867cf8a7b9acec442b249f000e0ff0b9e3dbb4c7c4c"} Dec 03 17:38:46 crc kubenswrapper[4787]: I1203 17:38:46.845062 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8331dfaa-fd3d-4609-a078-f16f781e0318","Type":"ContainerStarted","Data":"116bb24f741e472538e9a223e958833d09bb33cc4e019dd7e0d15fb2a82cad48"} Dec 03 17:38:46 crc kubenswrapper[4787]: I1203 17:38:46.850737 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4282e0c2-7846-4f1f-afeb-076726cfe29d","Type":"ContainerStarted","Data":"620f3781c6005f73c56e92517da3a7998413f902ca506cca186ebe2fb3525db8"} Dec 03 17:38:47 crc kubenswrapper[4787]: I1203 17:38:47.877413 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pwxs4" podUID="8d10b8a6-f576-452e-8fe1-14908ea925b3" containerName="registry-server" containerID="cri-o://18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03" gracePeriod=2 Dec 03 17:38:47 crc kubenswrapper[4787]: I1203 17:38:47.877881 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8331dfaa-fd3d-4609-a078-f16f781e0318","Type":"ContainerStarted","Data":"05fb673f47a606b1739278e9f5d76945330e35f5ea49b781a119745a2e9c534b"} Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.046579 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.739046 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.828047 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-catalog-content\") pod \"8d10b8a6-f576-452e-8fe1-14908ea925b3\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.828429 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-utilities\") pod \"8d10b8a6-f576-452e-8fe1-14908ea925b3\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.828651 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jftpb\" (UniqueName: \"kubernetes.io/projected/8d10b8a6-f576-452e-8fe1-14908ea925b3-kube-api-access-jftpb\") pod \"8d10b8a6-f576-452e-8fe1-14908ea925b3\" (UID: \"8d10b8a6-f576-452e-8fe1-14908ea925b3\") " Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.830502 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-utilities" (OuterVolumeSpecName: "utilities") pod "8d10b8a6-f576-452e-8fe1-14908ea925b3" (UID: "8d10b8a6-f576-452e-8fe1-14908ea925b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.837904 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d10b8a6-f576-452e-8fe1-14908ea925b3-kube-api-access-jftpb" (OuterVolumeSpecName: "kube-api-access-jftpb") pod "8d10b8a6-f576-452e-8fe1-14908ea925b3" (UID: "8d10b8a6-f576-452e-8fe1-14908ea925b3"). InnerVolumeSpecName "kube-api-access-jftpb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.858638 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8d10b8a6-f576-452e-8fe1-14908ea925b3" (UID: "8d10b8a6-f576-452e-8fe1-14908ea925b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.890650 4787 generic.go:334] "Generic (PLEG): container finished" podID="8d10b8a6-f576-452e-8fe1-14908ea925b3" containerID="18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03" exitCode=0 Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.890735 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pwxs4" Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.890752 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwxs4" event={"ID":"8d10b8a6-f576-452e-8fe1-14908ea925b3","Type":"ContainerDied","Data":"18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03"} Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.895600 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pwxs4" event={"ID":"8d10b8a6-f576-452e-8fe1-14908ea925b3","Type":"ContainerDied","Data":"0bc9a6f479f79dd9065e6ceb81a9cd91ae2fe055c7e635bdf578134b09742727"} Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.895636 4787 scope.go:117] "RemoveContainer" containerID="18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03" Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.906277 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4282e0c2-7846-4f1f-afeb-076726cfe29d","Type":"ContainerStarted","Data":"7ea84ed69f546d0057496b37bad2ff9e4768d95f9ba2190731bd6e284ffa7a63"} Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.906469 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-api" containerID="cri-o://fc7909877fcab48f36abd4f3e735bc3e386dede90fc29fb0260a7deda5ac65b4" gracePeriod=30 Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.906494 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-listener" containerID="cri-o://7ea84ed69f546d0057496b37bad2ff9e4768d95f9ba2190731bd6e284ffa7a63" gracePeriod=30 Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.906555 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-evaluator" containerID="cri-o://1adb2a520fce01177ef81401295860b153f5aab544859ed528f5654fa3326feb" gracePeriod=30 Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.906532 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-notifier" containerID="cri-o://620f3781c6005f73c56e92517da3a7998413f902ca506cca186ebe2fb3525db8" gracePeriod=30 Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.953994 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.718626702 podStartE2EDuration="11.953971017s" podCreationTimestamp="2025-12-03 17:38:37 +0000 UTC" firstStartedPulling="2025-12-03 17:38:39.013353599 +0000 UTC m=+1535.830824558" lastFinishedPulling="2025-12-03 17:38:48.248697904 +0000 UTC m=+1545.066168873" observedRunningTime="2025-12-03 17:38:48.944640834 +0000 UTC m=+1545.762111793" watchObservedRunningTime="2025-12-03 17:38:48.953971017 +0000 UTC m=+1545.771441976" Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.960330 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.960380 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jftpb\" (UniqueName: \"kubernetes.io/projected/8d10b8a6-f576-452e-8fe1-14908ea925b3-kube-api-access-jftpb\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.961242 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8d10b8a6-f576-452e-8fe1-14908ea925b3-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:48 crc kubenswrapper[4787]: I1203 17:38:48.996248 4787 scope.go:117] "RemoveContainer" containerID="57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a" Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.059304 4787 scope.go:117] "RemoveContainer" containerID="5576f8717aaeb4a23c47314f1af15835aa4f498a16dd1293132c8193609f8127" Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.068931 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pwxs4"] Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.084582 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pwxs4"] Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.102540 4787 scope.go:117] "RemoveContainer" containerID="18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03" Dec 03 17:38:49 crc kubenswrapper[4787]: E1203 17:38:49.105294 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03\": container with ID starting with 18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03 not found: ID does not exist" containerID="18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03" Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.105409 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03"} err="failed to get container status \"18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03\": rpc error: code = NotFound desc = could not find container \"18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03\": container with ID starting with 18ff0dd430546b69a1367a82a8c9dfdf4095ab146ffcd1701137306d65a95d03 not found: ID does not exist" Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.105496 4787 scope.go:117] "RemoveContainer" containerID="57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a" Dec 03 17:38:49 crc kubenswrapper[4787]: E1203 17:38:49.109696 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a\": container with ID starting with 57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a not found: ID does not exist" containerID="57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a" Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.109846 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a"} err="failed to get container status \"57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a\": rpc error: code = NotFound desc = could not find container \"57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a\": container with ID starting with 57bae2d2add854219b8f50220247a84c6bdd689f26da79397820a3bfee5f774a not found: ID does not exist" Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.109931 4787 scope.go:117] "RemoveContainer" containerID="5576f8717aaeb4a23c47314f1af15835aa4f498a16dd1293132c8193609f8127" Dec 03 17:38:49 crc kubenswrapper[4787]: E1203 17:38:49.114173 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5576f8717aaeb4a23c47314f1af15835aa4f498a16dd1293132c8193609f8127\": container with ID starting with 5576f8717aaeb4a23c47314f1af15835aa4f498a16dd1293132c8193609f8127 not found: ID does not exist" containerID="5576f8717aaeb4a23c47314f1af15835aa4f498a16dd1293132c8193609f8127" Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.114317 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5576f8717aaeb4a23c47314f1af15835aa4f498a16dd1293132c8193609f8127"} err="failed to get container status \"5576f8717aaeb4a23c47314f1af15835aa4f498a16dd1293132c8193609f8127\": rpc error: code = NotFound desc = could not find container \"5576f8717aaeb4a23c47314f1af15835aa4f498a16dd1293132c8193609f8127\": container with ID starting with 5576f8717aaeb4a23c47314f1af15835aa4f498a16dd1293132c8193609f8127 not found: ID does not exist" Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.781349 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d10b8a6-f576-452e-8fe1-14908ea925b3" path="/var/lib/kubelet/pods/8d10b8a6-f576-452e-8fe1-14908ea925b3/volumes" Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.924502 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8331dfaa-fd3d-4609-a078-f16f781e0318","Type":"ContainerStarted","Data":"b1deb8aa514910b7a189e732dec75d05f2892fafda65c91cbf2b0953b7674ba5"} Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.928187 4787 generic.go:334] "Generic (PLEG): container finished" podID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerID="1adb2a520fce01177ef81401295860b153f5aab544859ed528f5654fa3326feb" exitCode=0 Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.928235 4787 generic.go:334] "Generic (PLEG): container finished" podID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerID="fc7909877fcab48f36abd4f3e735bc3e386dede90fc29fb0260a7deda5ac65b4" exitCode=0 Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.928262 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4282e0c2-7846-4f1f-afeb-076726cfe29d","Type":"ContainerDied","Data":"1adb2a520fce01177ef81401295860b153f5aab544859ed528f5654fa3326feb"} Dec 03 17:38:49 crc kubenswrapper[4787]: I1203 17:38:49.928325 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4282e0c2-7846-4f1f-afeb-076726cfe29d","Type":"ContainerDied","Data":"fc7909877fcab48f36abd4f3e735bc3e386dede90fc29fb0260a7deda5ac65b4"} Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.679987 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.700147 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9jkn\" (UniqueName: \"kubernetes.io/projected/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-kube-api-access-v9jkn\") pod \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.700398 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-config-data\") pod \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.700507 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-combined-ca-bundle\") pod \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\" (UID: \"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339\") " Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.709239 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-kube-api-access-v9jkn" (OuterVolumeSpecName: "kube-api-access-v9jkn") pod "7b772ec2-ecb8-4c2c-8cff-a7aef1a65339" (UID: "7b772ec2-ecb8-4c2c-8cff-a7aef1a65339"). InnerVolumeSpecName "kube-api-access-v9jkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.753172 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-config-data" (OuterVolumeSpecName: "config-data") pod "7b772ec2-ecb8-4c2c-8cff-a7aef1a65339" (UID: "7b772ec2-ecb8-4c2c-8cff-a7aef1a65339"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.810411 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.810448 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9jkn\" (UniqueName: \"kubernetes.io/projected/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-kube-api-access-v9jkn\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.840200 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7b772ec2-ecb8-4c2c-8cff-a7aef1a65339" (UID: "7b772ec2-ecb8-4c2c-8cff-a7aef1a65339"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.912537 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.939751 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8331dfaa-fd3d-4609-a078-f16f781e0318","Type":"ContainerStarted","Data":"f45610d34420c8dc1e1a6cc010e0fd60c2c0c212c217cb1ef1e086ffadbd054d"} Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.939965 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="ceilometer-central-agent" containerID="cri-o://9f6f2df677b9f31d2bb45867cf8a7b9acec442b249f000e0ff0b9e3dbb4c7c4c" gracePeriod=30 Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.940272 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.940837 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="proxy-httpd" containerID="cri-o://f45610d34420c8dc1e1a6cc010e0fd60c2c0c212c217cb1ef1e086ffadbd054d" gracePeriod=30 Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.940895 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="sg-core" containerID="cri-o://b1deb8aa514910b7a189e732dec75d05f2892fafda65c91cbf2b0953b7674ba5" gracePeriod=30 Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.940935 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="ceilometer-notification-agent" containerID="cri-o://05fb673f47a606b1739278e9f5d76945330e35f5ea49b781a119745a2e9c534b" gracePeriod=30 Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.945710 4787 generic.go:334] "Generic (PLEG): container finished" podID="7b772ec2-ecb8-4c2c-8cff-a7aef1a65339" containerID="174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271" exitCode=137 Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.945757 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339","Type":"ContainerDied","Data":"174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271"} Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.945790 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7b772ec2-ecb8-4c2c-8cff-a7aef1a65339","Type":"ContainerDied","Data":"cf7d787e7c0464886741ac0a68896bbae42670876ce7d94d10988938763b074b"} Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.945811 4787 scope.go:117] "RemoveContainer" containerID="174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271" Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.945947 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:38:50 crc kubenswrapper[4787]: I1203 17:38:50.999251 4787 scope.go:117] "RemoveContainer" containerID="174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271" Dec 03 17:38:51 crc kubenswrapper[4787]: E1203 17:38:51.001874 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271\": container with ID starting with 174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271 not found: ID does not exist" containerID="174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.001914 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271"} err="failed to get container status \"174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271\": rpc error: code = NotFound desc = could not find container \"174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271\": container with ID starting with 174847914f478866792984323f4e573d5a0d6296611a1052669001bcfe430271 not found: ID does not exist" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.006847 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.847500254 podStartE2EDuration="7.006808649s" podCreationTimestamp="2025-12-03 17:38:44 +0000 UTC" firstStartedPulling="2025-12-03 17:38:45.967112971 +0000 UTC m=+1542.784583930" lastFinishedPulling="2025-12-03 17:38:50.126421366 +0000 UTC m=+1546.943892325" observedRunningTime="2025-12-03 17:38:50.977818984 +0000 UTC m=+1547.795289943" watchObservedRunningTime="2025-12-03 17:38:51.006808649 +0000 UTC m=+1547.824279638" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.032788 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.046566 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.061084 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:38:51 crc kubenswrapper[4787]: E1203 17:38:51.061592 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7b772ec2-ecb8-4c2c-8cff-a7aef1a65339" containerName="nova-scheduler-scheduler" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.061615 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7b772ec2-ecb8-4c2c-8cff-a7aef1a65339" containerName="nova-scheduler-scheduler" Dec 03 17:38:51 crc kubenswrapper[4787]: E1203 17:38:51.061642 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d10b8a6-f576-452e-8fe1-14908ea925b3" containerName="extract-utilities" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.061650 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d10b8a6-f576-452e-8fe1-14908ea925b3" containerName="extract-utilities" Dec 03 17:38:51 crc kubenswrapper[4787]: E1203 17:38:51.061687 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d10b8a6-f576-452e-8fe1-14908ea925b3" containerName="extract-content" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.061695 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d10b8a6-f576-452e-8fe1-14908ea925b3" containerName="extract-content" Dec 03 17:38:51 crc kubenswrapper[4787]: E1203 17:38:51.061717 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d10b8a6-f576-452e-8fe1-14908ea925b3" containerName="registry-server" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.061725 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d10b8a6-f576-452e-8fe1-14908ea925b3" containerName="registry-server" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.061983 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d10b8a6-f576-452e-8fe1-14908ea925b3" containerName="registry-server" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.062012 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7b772ec2-ecb8-4c2c-8cff-a7aef1a65339" containerName="nova-scheduler-scheduler" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.063101 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.065762 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.080221 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.116519 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.116664 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-config-data\") pod \"nova-scheduler-0\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.116704 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ptzwb\" (UniqueName: \"kubernetes.io/projected/c18af891-9fd6-48ea-a881-357223dbabbe-kube-api-access-ptzwb\") pod \"nova-scheduler-0\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.218778 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-config-data\") pod \"nova-scheduler-0\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.219107 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ptzwb\" (UniqueName: \"kubernetes.io/projected/c18af891-9fd6-48ea-a881-357223dbabbe-kube-api-access-ptzwb\") pod \"nova-scheduler-0\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.219242 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.222741 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-config-data\") pod \"nova-scheduler-0\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.223096 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.237237 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ptzwb\" (UniqueName: \"kubernetes.io/projected/c18af891-9fd6-48ea-a881-357223dbabbe-kube-api-access-ptzwb\") pod \"nova-scheduler-0\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " pod="openstack/nova-scheduler-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.390394 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.593146 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.630609 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjgz8\" (UniqueName: \"kubernetes.io/projected/b96e978f-7e73-45ea-9cdc-73b4744bee47-kube-api-access-cjgz8\") pod \"b96e978f-7e73-45ea-9cdc-73b4744bee47\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.630946 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-combined-ca-bundle\") pod \"b96e978f-7e73-45ea-9cdc-73b4744bee47\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.631244 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-config-data\") pod \"b96e978f-7e73-45ea-9cdc-73b4744bee47\" (UID: \"b96e978f-7e73-45ea-9cdc-73b4744bee47\") " Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.638148 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b96e978f-7e73-45ea-9cdc-73b4744bee47-kube-api-access-cjgz8" (OuterVolumeSpecName: "kube-api-access-cjgz8") pod "b96e978f-7e73-45ea-9cdc-73b4744bee47" (UID: "b96e978f-7e73-45ea-9cdc-73b4744bee47"). InnerVolumeSpecName "kube-api-access-cjgz8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.662971 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-config-data" (OuterVolumeSpecName: "config-data") pod "b96e978f-7e73-45ea-9cdc-73b4744bee47" (UID: "b96e978f-7e73-45ea-9cdc-73b4744bee47"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.676681 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b96e978f-7e73-45ea-9cdc-73b4744bee47" (UID: "b96e978f-7e73-45ea-9cdc-73b4744bee47"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.733469 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.733521 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b96e978f-7e73-45ea-9cdc-73b4744bee47-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.733535 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjgz8\" (UniqueName: \"kubernetes.io/projected/b96e978f-7e73-45ea-9cdc-73b4744bee47-kube-api-access-cjgz8\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.778381 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7b772ec2-ecb8-4c2c-8cff-a7aef1a65339" path="/var/lib/kubelet/pods/7b772ec2-ecb8-4c2c-8cff-a7aef1a65339/volumes" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.901244 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:38:51 crc kubenswrapper[4787]: W1203 17:38:51.902230 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc18af891_9fd6_48ea_a881_357223dbabbe.slice/crio-53e508d9d1d7fe81a8de542fb46d7c9c99e3491eb4e463ad852189f0191a86ae WatchSource:0}: Error finding container 53e508d9d1d7fe81a8de542fb46d7c9c99e3491eb4e463ad852189f0191a86ae: Status 404 returned error can't find the container with id 53e508d9d1d7fe81a8de542fb46d7c9c99e3491eb4e463ad852189f0191a86ae Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.959586 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c18af891-9fd6-48ea-a881-357223dbabbe","Type":"ContainerStarted","Data":"53e508d9d1d7fe81a8de542fb46d7c9c99e3491eb4e463ad852189f0191a86ae"} Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.961348 4787 generic.go:334] "Generic (PLEG): container finished" podID="b96e978f-7e73-45ea-9cdc-73b4744bee47" containerID="8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354" exitCode=137 Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.961416 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b96e978f-7e73-45ea-9cdc-73b4744bee47","Type":"ContainerDied","Data":"8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354"} Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.961428 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.961462 4787 scope.go:117] "RemoveContainer" containerID="8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354" Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.961449 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"b96e978f-7e73-45ea-9cdc-73b4744bee47","Type":"ContainerDied","Data":"3a159799acec866dc5792a10af3e986a3e6cc8ae6b545e508296c59852983ff8"} Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.966810 4787 generic.go:334] "Generic (PLEG): container finished" podID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerID="f45610d34420c8dc1e1a6cc010e0fd60c2c0c212c217cb1ef1e086ffadbd054d" exitCode=0 Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.966844 4787 generic.go:334] "Generic (PLEG): container finished" podID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerID="b1deb8aa514910b7a189e732dec75d05f2892fafda65c91cbf2b0953b7674ba5" exitCode=2 Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.966852 4787 generic.go:334] "Generic (PLEG): container finished" podID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerID="05fb673f47a606b1739278e9f5d76945330e35f5ea49b781a119745a2e9c534b" exitCode=0 Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.966872 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8331dfaa-fd3d-4609-a078-f16f781e0318","Type":"ContainerDied","Data":"f45610d34420c8dc1e1a6cc010e0fd60c2c0c212c217cb1ef1e086ffadbd054d"} Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.966941 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8331dfaa-fd3d-4609-a078-f16f781e0318","Type":"ContainerDied","Data":"b1deb8aa514910b7a189e732dec75d05f2892fafda65c91cbf2b0953b7674ba5"} Dec 03 17:38:51 crc kubenswrapper[4787]: I1203 17:38:51.966956 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8331dfaa-fd3d-4609-a078-f16f781e0318","Type":"ContainerDied","Data":"05fb673f47a606b1739278e9f5d76945330e35f5ea49b781a119745a2e9c534b"} Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.008048 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.009738 4787 scope.go:117] "RemoveContainer" containerID="8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354" Dec 03 17:38:52 crc kubenswrapper[4787]: E1203 17:38:52.010362 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354\": container with ID starting with 8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354 not found: ID does not exist" containerID="8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.010394 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354"} err="failed to get container status \"8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354\": rpc error: code = NotFound desc = could not find container \"8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354\": container with ID starting with 8359df3970ca702e90e3a28b16c7bb0ce6de53950bacc1a99abe33b0df67a354 not found: ID does not exist" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.028692 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.043907 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:38:52 crc kubenswrapper[4787]: E1203 17:38:52.044960 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b96e978f-7e73-45ea-9cdc-73b4744bee47" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.044988 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b96e978f-7e73-45ea-9cdc-73b4744bee47" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.060165 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b96e978f-7e73-45ea-9cdc-73b4744bee47" containerName="nova-cell1-novncproxy-novncproxy" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.061675 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.065586 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.065797 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.065895 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.074971 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.150311 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.150374 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.150451 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.150475 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-529j8\" (UniqueName: \"kubernetes.io/projected/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-kube-api-access-529j8\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.150503 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.251977 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.252058 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.252140 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.252165 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-529j8\" (UniqueName: \"kubernetes.io/projected/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-kube-api-access-529j8\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.252195 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.256275 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.256614 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.258069 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.258522 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.277493 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-529j8\" (UniqueName: \"kubernetes.io/projected/47dc3b7d-f87f-4e03-b969-242a9fdbee2b-kube-api-access-529j8\") pod \"nova-cell1-novncproxy-0\" (UID: \"47dc3b7d-f87f-4e03-b969-242a9fdbee2b\") " pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.382952 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.909536 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 03 17:38:52 crc kubenswrapper[4787]: W1203 17:38:52.909611 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod47dc3b7d_f87f_4e03_b969_242a9fdbee2b.slice/crio-7d2be81d889f2f0d0b278c977e5952e41f009bdd88564196495b3840a91ce233 WatchSource:0}: Error finding container 7d2be81d889f2f0d0b278c977e5952e41f009bdd88564196495b3840a91ce233: Status 404 returned error can't find the container with id 7d2be81d889f2f0d0b278c977e5952e41f009bdd88564196495b3840a91ce233 Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.984689 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c18af891-9fd6-48ea-a881-357223dbabbe","Type":"ContainerStarted","Data":"3b0937001aa222d83f9c5e71a21aae35441a160c418fc4e451e54d73c16f2796"} Dec 03 17:38:52 crc kubenswrapper[4787]: I1203 17:38:52.987629 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"47dc3b7d-f87f-4e03-b969-242a9fdbee2b","Type":"ContainerStarted","Data":"7d2be81d889f2f0d0b278c977e5952e41f009bdd88564196495b3840a91ce233"} Dec 03 17:38:53 crc kubenswrapper[4787]: I1203 17:38:53.010804 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.010754536 podStartE2EDuration="3.010754536s" podCreationTimestamp="2025-12-03 17:38:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:53.005157665 +0000 UTC m=+1549.822628624" watchObservedRunningTime="2025-12-03 17:38:53.010754536 +0000 UTC m=+1549.828225495" Dec 03 17:38:53 crc kubenswrapper[4787]: I1203 17:38:53.807912 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b96e978f-7e73-45ea-9cdc-73b4744bee47" path="/var/lib/kubelet/pods/b96e978f-7e73-45ea-9cdc-73b4744bee47/volumes" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.049413 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"47dc3b7d-f87f-4e03-b969-242a9fdbee2b","Type":"ContainerStarted","Data":"3332d7a8648d5be4c43e5fc37f8b230f4ddcda8f894efce2cb62174c0a5f0046"} Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.095389 4787 generic.go:334] "Generic (PLEG): container finished" podID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerID="9f6f2df677b9f31d2bb45867cf8a7b9acec442b249f000e0ff0b9e3dbb4c7c4c" exitCode=0 Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.096129 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8331dfaa-fd3d-4609-a078-f16f781e0318","Type":"ContainerDied","Data":"9f6f2df677b9f31d2bb45867cf8a7b9acec442b249f000e0ff0b9e3dbb4c7c4c"} Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.120232 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.120213971 podStartE2EDuration="2.120213971s" podCreationTimestamp="2025-12-03 17:38:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:54.082567951 +0000 UTC m=+1550.900038910" watchObservedRunningTime="2025-12-03 17:38:54.120213971 +0000 UTC m=+1550.937684930" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.295716 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.428696 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-combined-ca-bundle\") pod \"8331dfaa-fd3d-4609-a078-f16f781e0318\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.428820 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkzjv\" (UniqueName: \"kubernetes.io/projected/8331dfaa-fd3d-4609-a078-f16f781e0318-kube-api-access-zkzjv\") pod \"8331dfaa-fd3d-4609-a078-f16f781e0318\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.428871 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-config-data\") pod \"8331dfaa-fd3d-4609-a078-f16f781e0318\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.428901 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-sg-core-conf-yaml\") pod \"8331dfaa-fd3d-4609-a078-f16f781e0318\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.428926 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-log-httpd\") pod \"8331dfaa-fd3d-4609-a078-f16f781e0318\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.429050 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-scripts\") pod \"8331dfaa-fd3d-4609-a078-f16f781e0318\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.429161 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-run-httpd\") pod \"8331dfaa-fd3d-4609-a078-f16f781e0318\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.429349 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8331dfaa-fd3d-4609-a078-f16f781e0318" (UID: "8331dfaa-fd3d-4609-a078-f16f781e0318"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.429587 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.429791 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8331dfaa-fd3d-4609-a078-f16f781e0318" (UID: "8331dfaa-fd3d-4609-a078-f16f781e0318"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.434190 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8331dfaa-fd3d-4609-a078-f16f781e0318-kube-api-access-zkzjv" (OuterVolumeSpecName: "kube-api-access-zkzjv") pod "8331dfaa-fd3d-4609-a078-f16f781e0318" (UID: "8331dfaa-fd3d-4609-a078-f16f781e0318"). InnerVolumeSpecName "kube-api-access-zkzjv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.435731 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-scripts" (OuterVolumeSpecName: "scripts") pod "8331dfaa-fd3d-4609-a078-f16f781e0318" (UID: "8331dfaa-fd3d-4609-a078-f16f781e0318"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.468837 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8331dfaa-fd3d-4609-a078-f16f781e0318" (UID: "8331dfaa-fd3d-4609-a078-f16f781e0318"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.530222 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8331dfaa-fd3d-4609-a078-f16f781e0318" (UID: "8331dfaa-fd3d-4609-a078-f16f781e0318"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.530428 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-combined-ca-bundle\") pod \"8331dfaa-fd3d-4609-a078-f16f781e0318\" (UID: \"8331dfaa-fd3d-4609-a078-f16f781e0318\") " Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.530989 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkzjv\" (UniqueName: \"kubernetes.io/projected/8331dfaa-fd3d-4609-a078-f16f781e0318-kube-api-access-zkzjv\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.531001 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.531010 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.531030 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8331dfaa-fd3d-4609-a078-f16f781e0318-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:54 crc kubenswrapper[4787]: W1203 17:38:54.531095 4787 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/8331dfaa-fd3d-4609-a078-f16f781e0318/volumes/kubernetes.io~secret/combined-ca-bundle Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.531106 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8331dfaa-fd3d-4609-a078-f16f781e0318" (UID: "8331dfaa-fd3d-4609-a078-f16f781e0318"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.544453 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-config-data" (OuterVolumeSpecName: "config-data") pod "8331dfaa-fd3d-4609-a078-f16f781e0318" (UID: "8331dfaa-fd3d-4609-a078-f16f781e0318"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.632445 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:54 crc kubenswrapper[4787]: I1203 17:38:54.632473 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8331dfaa-fd3d-4609-a078-f16f781e0318-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.014111 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.015787 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.022362 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.087598 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.087968 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.088313 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.090385 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.111720 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8331dfaa-fd3d-4609-a078-f16f781e0318","Type":"ContainerDied","Data":"116bb24f741e472538e9a223e958833d09bb33cc4e019dd7e0d15fb2a82cad48"} Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.111784 4787 scope.go:117] "RemoveContainer" containerID="f45610d34420c8dc1e1a6cc010e0fd60c2c0c212c217cb1ef1e086ffadbd054d" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.112077 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.112527 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.116243 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.118354 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.141217 4787 scope.go:117] "RemoveContainer" containerID="b1deb8aa514910b7a189e732dec75d05f2892fafda65c91cbf2b0953b7674ba5" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.196551 4787 scope.go:117] "RemoveContainer" containerID="05fb673f47a606b1739278e9f5d76945330e35f5ea49b781a119745a2e9c534b" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.205516 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.210746 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.235284 4787 scope.go:117] "RemoveContainer" containerID="9f6f2df677b9f31d2bb45867cf8a7b9acec442b249f000e0ff0b9e3dbb4c7c4c" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.254079 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:55 crc kubenswrapper[4787]: E1203 17:38:55.254520 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="proxy-httpd" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.254539 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="proxy-httpd" Dec 03 17:38:55 crc kubenswrapper[4787]: E1203 17:38:55.254562 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="ceilometer-notification-agent" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.254568 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="ceilometer-notification-agent" Dec 03 17:38:55 crc kubenswrapper[4787]: E1203 17:38:55.254582 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="ceilometer-central-agent" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.254588 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="ceilometer-central-agent" Dec 03 17:38:55 crc kubenswrapper[4787]: E1203 17:38:55.254612 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="sg-core" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.254620 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="sg-core" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.254813 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="proxy-httpd" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.254831 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="ceilometer-central-agent" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.254846 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="ceilometer-notification-agent" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.254855 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" containerName="sg-core" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.258202 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.268404 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.269287 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.319093 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.368910 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-whzrl"] Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.371385 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.374072 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4nwj\" (UniqueName: \"kubernetes.io/projected/824eb81c-fae7-4457-a6ac-88e8197d6bc2-kube-api-access-p4nwj\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.374124 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-run-httpd\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.374159 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-config-data\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.374279 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.374304 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-scripts\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.374354 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-log-httpd\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.374540 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.393087 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-whzrl"] Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.445176 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-bs2bm"] Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.447416 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.455126 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bs2bm"] Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.476916 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477036 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477078 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-scripts\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477128 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-log-httpd\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477170 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477225 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-config\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477245 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477322 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477384 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4nwj\" (UniqueName: \"kubernetes.io/projected/824eb81c-fae7-4457-a6ac-88e8197d6bc2-kube-api-access-p4nwj\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477405 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-run-httpd\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477433 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-config-data\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477458 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.477490 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gtlpz\" (UniqueName: \"kubernetes.io/projected/9ac14488-1fe6-458a-a812-e4b62016f67f-kube-api-access-gtlpz\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.478777 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-log-httpd\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.480951 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-run-httpd\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.489310 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.489410 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.490422 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-scripts\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.498698 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4nwj\" (UniqueName: \"kubernetes.io/projected/824eb81c-fae7-4457-a6ac-88e8197d6bc2-kube-api-access-p4nwj\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.501284 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-config-data\") pod \"ceilometer-0\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.579873 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-catalog-content\") pod \"community-operators-bs2bm\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.579957 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.580116 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4qzsn\" (UniqueName: \"kubernetes.io/projected/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-kube-api-access-4qzsn\") pod \"community-operators-bs2bm\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.580143 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.585318 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-config\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.585358 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.581874 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-nb\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.585619 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.585672 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gtlpz\" (UniqueName: \"kubernetes.io/projected/9ac14488-1fe6-458a-a812-e4b62016f67f-kube-api-access-gtlpz\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.585786 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-utilities\") pod \"community-operators-bs2bm\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.581233 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-svc\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.586528 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-config\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.586648 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-swift-storage-0\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.586797 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-sb\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.606003 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gtlpz\" (UniqueName: \"kubernetes.io/projected/9ac14488-1fe6-458a-a812-e4b62016f67f-kube-api-access-gtlpz\") pod \"dnsmasq-dns-6b7bbf7cf9-whzrl\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.607631 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.687422 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-utilities\") pod \"community-operators-bs2bm\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.687738 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-catalog-content\") pod \"community-operators-bs2bm\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.687837 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4qzsn\" (UniqueName: \"kubernetes.io/projected/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-kube-api-access-4qzsn\") pod \"community-operators-bs2bm\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.688050 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-utilities\") pod \"community-operators-bs2bm\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.688275 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-catalog-content\") pod \"community-operators-bs2bm\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.699974 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.708528 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4qzsn\" (UniqueName: \"kubernetes.io/projected/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-kube-api-access-4qzsn\") pod \"community-operators-bs2bm\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.790667 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:38:55 crc kubenswrapper[4787]: I1203 17:38:55.790693 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8331dfaa-fd3d-4609-a078-f16f781e0318" path="/var/lib/kubelet/pods/8331dfaa-fd3d-4609-a078-f16f781e0318/volumes" Dec 03 17:38:56 crc kubenswrapper[4787]: W1203 17:38:56.189510 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod824eb81c_fae7_4457_a6ac_88e8197d6bc2.slice/crio-c9d3968f8272a2653916edea12140105d9eaecf96bf52184812426e07f56363f WatchSource:0}: Error finding container c9d3968f8272a2653916edea12140105d9eaecf96bf52184812426e07f56363f: Status 404 returned error can't find the container with id c9d3968f8272a2653916edea12140105d9eaecf96bf52184812426e07f56363f Dec 03 17:38:56 crc kubenswrapper[4787]: I1203 17:38:56.194176 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:56 crc kubenswrapper[4787]: I1203 17:38:56.391449 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 17:38:56 crc kubenswrapper[4787]: W1203 17:38:56.414761 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ac14488_1fe6_458a_a812_e4b62016f67f.slice/crio-6ffd935f48feb5a51bde831cf97ce2cdeed11a4270d24095a1cba5af42864dfc WatchSource:0}: Error finding container 6ffd935f48feb5a51bde831cf97ce2cdeed11a4270d24095a1cba5af42864dfc: Status 404 returned error can't find the container with id 6ffd935f48feb5a51bde831cf97ce2cdeed11a4270d24095a1cba5af42864dfc Dec 03 17:38:56 crc kubenswrapper[4787]: I1203 17:38:56.419306 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-whzrl"] Dec 03 17:38:56 crc kubenswrapper[4787]: I1203 17:38:56.450768 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-bs2bm"] Dec 03 17:38:57 crc kubenswrapper[4787]: I1203 17:38:57.187812 4787 generic.go:334] "Generic (PLEG): container finished" podID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" containerID="230cf3cfeebe79f85db9a716451a99978d9a8fdcb219c52d6ffd18da5310f7e8" exitCode=0 Dec 03 17:38:57 crc kubenswrapper[4787]: I1203 17:38:57.188122 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs2bm" event={"ID":"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae","Type":"ContainerDied","Data":"230cf3cfeebe79f85db9a716451a99978d9a8fdcb219c52d6ffd18da5310f7e8"} Dec 03 17:38:57 crc kubenswrapper[4787]: I1203 17:38:57.188473 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs2bm" event={"ID":"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae","Type":"ContainerStarted","Data":"8e53dba77da2c4af6ea54e266c59e670d94070c8a9d4c3288028485832dd9bc5"} Dec 03 17:38:57 crc kubenswrapper[4787]: I1203 17:38:57.192583 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" event={"ID":"9ac14488-1fe6-458a-a812-e4b62016f67f","Type":"ContainerStarted","Data":"8ef8ae651004d83911726f1a859674c0eaea06e6d301fb5285a3bbcf04b6e519"} Dec 03 17:38:57 crc kubenswrapper[4787]: I1203 17:38:57.193254 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" event={"ID":"9ac14488-1fe6-458a-a812-e4b62016f67f","Type":"ContainerStarted","Data":"6ffd935f48feb5a51bde831cf97ce2cdeed11a4270d24095a1cba5af42864dfc"} Dec 03 17:38:57 crc kubenswrapper[4787]: I1203 17:38:57.200549 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"824eb81c-fae7-4457-a6ac-88e8197d6bc2","Type":"ContainerStarted","Data":"e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6"} Dec 03 17:38:57 crc kubenswrapper[4787]: I1203 17:38:57.200594 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"824eb81c-fae7-4457-a6ac-88e8197d6bc2","Type":"ContainerStarted","Data":"c9d3968f8272a2653916edea12140105d9eaecf96bf52184812426e07f56363f"} Dec 03 17:38:57 crc kubenswrapper[4787]: I1203 17:38:57.383255 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:38:58 crc kubenswrapper[4787]: I1203 17:38:58.214336 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"824eb81c-fae7-4457-a6ac-88e8197d6bc2","Type":"ContainerStarted","Data":"fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e"} Dec 03 17:38:58 crc kubenswrapper[4787]: I1203 17:38:58.218586 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs2bm" event={"ID":"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae","Type":"ContainerStarted","Data":"f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399"} Dec 03 17:38:58 crc kubenswrapper[4787]: I1203 17:38:58.222754 4787 generic.go:334] "Generic (PLEG): container finished" podID="9ac14488-1fe6-458a-a812-e4b62016f67f" containerID="8ef8ae651004d83911726f1a859674c0eaea06e6d301fb5285a3bbcf04b6e519" exitCode=0 Dec 03 17:38:58 crc kubenswrapper[4787]: I1203 17:38:58.222827 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" event={"ID":"9ac14488-1fe6-458a-a812-e4b62016f67f","Type":"ContainerDied","Data":"8ef8ae651004d83911726f1a859674c0eaea06e6d301fb5285a3bbcf04b6e519"} Dec 03 17:38:58 crc kubenswrapper[4787]: I1203 17:38:58.222857 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" event={"ID":"9ac14488-1fe6-458a-a812-e4b62016f67f","Type":"ContainerStarted","Data":"2669ec334559ba3f1ee8d39ff64ee6f2897545548e3904e7bb411fee93de4aae"} Dec 03 17:38:58 crc kubenswrapper[4787]: I1203 17:38:58.223771 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:38:58 crc kubenswrapper[4787]: I1203 17:38:58.275882 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" podStartSLOduration=3.275862086 podStartE2EDuration="3.275862086s" podCreationTimestamp="2025-12-03 17:38:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:38:58.272432154 +0000 UTC m=+1555.089903123" watchObservedRunningTime="2025-12-03 17:38:58.275862086 +0000 UTC m=+1555.093333045" Dec 03 17:38:58 crc kubenswrapper[4787]: I1203 17:38:58.494906 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:38:58 crc kubenswrapper[4787]: I1203 17:38:58.495371 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerName="nova-api-log" containerID="cri-o://8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499" gracePeriod=30 Dec 03 17:38:58 crc kubenswrapper[4787]: I1203 17:38:58.495459 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerName="nova-api-api" containerID="cri-o://df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d" gracePeriod=30 Dec 03 17:38:58 crc kubenswrapper[4787]: I1203 17:38:58.743418 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:38:59 crc kubenswrapper[4787]: I1203 17:38:59.241090 4787 generic.go:334] "Generic (PLEG): container finished" podID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" containerID="f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399" exitCode=0 Dec 03 17:38:59 crc kubenswrapper[4787]: I1203 17:38:59.241193 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs2bm" event={"ID":"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae","Type":"ContainerDied","Data":"f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399"} Dec 03 17:39:00 crc kubenswrapper[4787]: I1203 17:39:00.253123 4787 generic.go:334] "Generic (PLEG): container finished" podID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerID="8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499" exitCode=143 Dec 03 17:39:00 crc kubenswrapper[4787]: I1203 17:39:00.253140 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d4b067d-cfb3-4923-ae34-0cdee637eb81","Type":"ContainerDied","Data":"8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499"} Dec 03 17:39:00 crc kubenswrapper[4787]: I1203 17:39:00.257614 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"824eb81c-fae7-4457-a6ac-88e8197d6bc2","Type":"ContainerStarted","Data":"6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1"} Dec 03 17:39:01 crc kubenswrapper[4787]: I1203 17:39:01.300133 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs2bm" event={"ID":"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae","Type":"ContainerStarted","Data":"90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec"} Dec 03 17:39:01 crc kubenswrapper[4787]: I1203 17:39:01.324774 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-bs2bm" podStartSLOduration=3.88832784 podStartE2EDuration="6.324754935s" podCreationTimestamp="2025-12-03 17:38:55 +0000 UTC" firstStartedPulling="2025-12-03 17:38:57.189478377 +0000 UTC m=+1554.006949346" lastFinishedPulling="2025-12-03 17:38:59.625905482 +0000 UTC m=+1556.443376441" observedRunningTime="2025-12-03 17:39:01.322103443 +0000 UTC m=+1558.139574402" watchObservedRunningTime="2025-12-03 17:39:01.324754935 +0000 UTC m=+1558.142225894" Dec 03 17:39:01 crc kubenswrapper[4787]: I1203 17:39:01.390969 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 17:39:01 crc kubenswrapper[4787]: I1203 17:39:01.447736 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.313783 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"824eb81c-fae7-4457-a6ac-88e8197d6bc2","Type":"ContainerStarted","Data":"7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1"} Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.315081 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.314885 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="proxy-httpd" containerID="cri-o://7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1" gracePeriod=30 Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.314903 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="sg-core" containerID="cri-o://6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1" gracePeriod=30 Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.314913 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="ceilometer-notification-agent" containerID="cri-o://fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e" gracePeriod=30 Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.314559 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="ceilometer-central-agent" containerID="cri-o://e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6" gracePeriod=30 Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.315749 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.318973 4787 generic.go:334] "Generic (PLEG): container finished" podID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerID="df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d" exitCode=0 Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.319106 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d4b067d-cfb3-4923-ae34-0cdee637eb81","Type":"ContainerDied","Data":"df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d"} Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.319168 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2d4b067d-cfb3-4923-ae34-0cdee637eb81","Type":"ContainerDied","Data":"e2f1198476ba2af04e100ad194b3485679b9e7e8d999da6216107837905a9563"} Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.319188 4787 scope.go:117] "RemoveContainer" containerID="df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.351246 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.339517788 podStartE2EDuration="7.351226784s" podCreationTimestamp="2025-12-03 17:38:55 +0000 UTC" firstStartedPulling="2025-12-03 17:38:56.193800903 +0000 UTC m=+1553.011271862" lastFinishedPulling="2025-12-03 17:39:01.205509899 +0000 UTC m=+1558.022980858" observedRunningTime="2025-12-03 17:39:02.343119494 +0000 UTC m=+1559.160590453" watchObservedRunningTime="2025-12-03 17:39:02.351226784 +0000 UTC m=+1559.168697743" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.355418 4787 scope.go:117] "RemoveContainer" containerID="8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.369983 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.383674 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.404043 4787 scope.go:117] "RemoveContainer" containerID="df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d" Dec 03 17:39:02 crc kubenswrapper[4787]: E1203 17:39:02.411291 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d\": container with ID starting with df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d not found: ID does not exist" containerID="df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.411356 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d"} err="failed to get container status \"df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d\": rpc error: code = NotFound desc = could not find container \"df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d\": container with ID starting with df40e402e9c53c54dd005f35f3859f4184ff8924742badadc11c5a4c4145b28d not found: ID does not exist" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.411381 4787 scope.go:117] "RemoveContainer" containerID="8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499" Dec 03 17:39:02 crc kubenswrapper[4787]: E1203 17:39:02.415542 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499\": container with ID starting with 8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499 not found: ID does not exist" containerID="8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.415581 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499"} err="failed to get container status \"8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499\": rpc error: code = NotFound desc = could not find container \"8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499\": container with ID starting with 8807eb8564ff38cbb1599490fb3b42dc6869ac4d4fe86ac4e5cea3c981c05499 not found: ID does not exist" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.476766 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.497065 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d4b067d-cfb3-4923-ae34-0cdee637eb81-logs\") pod \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.497144 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-shd4h\" (UniqueName: \"kubernetes.io/projected/2d4b067d-cfb3-4923-ae34-0cdee637eb81-kube-api-access-shd4h\") pod \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.497179 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-config-data\") pod \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.497240 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-combined-ca-bundle\") pod \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\" (UID: \"2d4b067d-cfb3-4923-ae34-0cdee637eb81\") " Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.499369 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2d4b067d-cfb3-4923-ae34-0cdee637eb81-logs" (OuterVolumeSpecName: "logs") pod "2d4b067d-cfb3-4923-ae34-0cdee637eb81" (UID: "2d4b067d-cfb3-4923-ae34-0cdee637eb81"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.522350 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d4b067d-cfb3-4923-ae34-0cdee637eb81-kube-api-access-shd4h" (OuterVolumeSpecName: "kube-api-access-shd4h") pod "2d4b067d-cfb3-4923-ae34-0cdee637eb81" (UID: "2d4b067d-cfb3-4923-ae34-0cdee637eb81"). InnerVolumeSpecName "kube-api-access-shd4h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.544614 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-config-data" (OuterVolumeSpecName: "config-data") pod "2d4b067d-cfb3-4923-ae34-0cdee637eb81" (UID: "2d4b067d-cfb3-4923-ae34-0cdee637eb81"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.552279 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2d4b067d-cfb3-4923-ae34-0cdee637eb81" (UID: "2d4b067d-cfb3-4923-ae34-0cdee637eb81"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.599827 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2d4b067d-cfb3-4923-ae34-0cdee637eb81-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.600234 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-shd4h\" (UniqueName: \"kubernetes.io/projected/2d4b067d-cfb3-4923-ae34-0cdee637eb81-kube-api-access-shd4h\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.600250 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:02 crc kubenswrapper[4787]: I1203 17:39:02.600264 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2d4b067d-cfb3-4923-ae34-0cdee637eb81-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.329833 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.332672 4787 generic.go:334] "Generic (PLEG): container finished" podID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerID="7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1" exitCode=0 Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.332711 4787 generic.go:334] "Generic (PLEG): container finished" podID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerID="6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1" exitCode=2 Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.332723 4787 generic.go:334] "Generic (PLEG): container finished" podID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerID="fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e" exitCode=0 Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.333914 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"824eb81c-fae7-4457-a6ac-88e8197d6bc2","Type":"ContainerDied","Data":"7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1"} Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.333946 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"824eb81c-fae7-4457-a6ac-88e8197d6bc2","Type":"ContainerDied","Data":"6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1"} Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.333973 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"824eb81c-fae7-4457-a6ac-88e8197d6bc2","Type":"ContainerDied","Data":"fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e"} Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.356813 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.362644 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.375728 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.391408 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 17:39:03 crc kubenswrapper[4787]: E1203 17:39:03.392131 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerName="nova-api-api" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.392238 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerName="nova-api-api" Dec 03 17:39:03 crc kubenswrapper[4787]: E1203 17:39:03.392355 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerName="nova-api-log" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.392416 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerName="nova-api-log" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.392712 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerName="nova-api-log" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.392778 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" containerName="nova-api-api" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.394342 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.397041 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.398833 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.399381 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.415257 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.520198 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-public-tls-certs\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.520333 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29wzw\" (UniqueName: \"kubernetes.io/projected/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-kube-api-access-29wzw\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.520378 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-logs\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.520426 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.520464 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-config-data\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.520520 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.580148 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-7gzqf"] Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.581682 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.584611 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.584717 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.596659 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7gzqf"] Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.622585 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29wzw\" (UniqueName: \"kubernetes.io/projected/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-kube-api-access-29wzw\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.622671 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-logs\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.622737 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.622777 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-config-data\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.622858 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.622945 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-public-tls-certs\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.623312 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-logs\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.628735 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-public-tls-certs\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.629400 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.629842 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-internal-tls-certs\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.638722 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-config-data\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.640542 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29wzw\" (UniqueName: \"kubernetes.io/projected/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-kube-api-access-29wzw\") pod \"nova-api-0\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.713592 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.724446 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-scripts\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.724558 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-config-data\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.724582 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bzx5d\" (UniqueName: \"kubernetes.io/projected/a28737aa-7ebe-4dbe-967d-51376202958e-kube-api-access-bzx5d\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.724633 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.786965 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2d4b067d-cfb3-4923-ae34-0cdee637eb81" path="/var/lib/kubelet/pods/2d4b067d-cfb3-4923-ae34-0cdee637eb81/volumes" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.826812 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.826965 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-scripts\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.827128 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-config-data\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.827176 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bzx5d\" (UniqueName: \"kubernetes.io/projected/a28737aa-7ebe-4dbe-967d-51376202958e-kube-api-access-bzx5d\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.835412 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.835597 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.841961 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-config-data\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.843550 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.849042 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-scripts\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.851855 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bzx5d\" (UniqueName: \"kubernetes.io/projected/a28737aa-7ebe-4dbe-967d-51376202958e-kube-api-access-bzx5d\") pod \"nova-cell1-cell-mapping-7gzqf\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:03 crc kubenswrapper[4787]: I1203 17:39:03.899997 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:04 crc kubenswrapper[4787]: I1203 17:39:04.255935 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:39:04 crc kubenswrapper[4787]: I1203 17:39:04.344775 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e","Type":"ContainerStarted","Data":"3de0b9dd110f11333a6f2378b94970c8e1660915e9e777f8bb4d38f2f7863fc6"} Dec 03 17:39:04 crc kubenswrapper[4787]: I1203 17:39:04.434892 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-7gzqf"] Dec 03 17:39:05 crc kubenswrapper[4787]: I1203 17:39:05.361580 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7gzqf" event={"ID":"a28737aa-7ebe-4dbe-967d-51376202958e","Type":"ContainerStarted","Data":"1a00a16219839ac3284d87eacc77080bb5386e413634aaad555e7c56871e39f0"} Dec 03 17:39:05 crc kubenswrapper[4787]: I1203 17:39:05.362087 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7gzqf" event={"ID":"a28737aa-7ebe-4dbe-967d-51376202958e","Type":"ContainerStarted","Data":"0f160cc92d9cbc35f7e03821298f905df77500da64d0e2b6ffa5f3ca462bbb79"} Dec 03 17:39:05 crc kubenswrapper[4787]: I1203 17:39:05.365058 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e","Type":"ContainerStarted","Data":"0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b"} Dec 03 17:39:05 crc kubenswrapper[4787]: I1203 17:39:05.365115 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e","Type":"ContainerStarted","Data":"6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430"} Dec 03 17:39:05 crc kubenswrapper[4787]: I1203 17:39:05.392438 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-7gzqf" podStartSLOduration=2.3924161870000002 podStartE2EDuration="2.392416187s" podCreationTimestamp="2025-12-03 17:39:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:39:05.379288451 +0000 UTC m=+1562.196759410" watchObservedRunningTime="2025-12-03 17:39:05.392416187 +0000 UTC m=+1562.209887146" Dec 03 17:39:05 crc kubenswrapper[4787]: I1203 17:39:05.411122 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.411101154 podStartE2EDuration="2.411101154s" podCreationTimestamp="2025-12-03 17:39:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:39:05.397428133 +0000 UTC m=+1562.214899172" watchObservedRunningTime="2025-12-03 17:39:05.411101154 +0000 UTC m=+1562.228572113" Dec 03 17:39:05 crc kubenswrapper[4787]: I1203 17:39:05.704236 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:39:05 crc kubenswrapper[4787]: I1203 17:39:05.791009 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:39:05 crc kubenswrapper[4787]: I1203 17:39:05.791126 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:39:05 crc kubenswrapper[4787]: I1203 17:39:05.800675 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6w5hx"] Dec 03 17:39:05 crc kubenswrapper[4787]: I1203 17:39:05.800932 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" podUID="c19a0911-5c3e-4c36-96ec-064cd73b8bde" containerName="dnsmasq-dns" containerID="cri-o://956351eadf8c23b4abdbbfb3c30e61772f4c35e4ebe96bdb403d10a529f50e2d" gracePeriod=10 Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.027972 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.254418 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.293298 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-config-data\") pod \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.293379 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-run-httpd\") pod \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.293431 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-sg-core-conf-yaml\") pod \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.293567 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-scripts\") pod \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.293591 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4nwj\" (UniqueName: \"kubernetes.io/projected/824eb81c-fae7-4457-a6ac-88e8197d6bc2-kube-api-access-p4nwj\") pod \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.293611 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-combined-ca-bundle\") pod \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.293653 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-log-httpd\") pod \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\" (UID: \"824eb81c-fae7-4457-a6ac-88e8197d6bc2\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.295480 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "824eb81c-fae7-4457-a6ac-88e8197d6bc2" (UID: "824eb81c-fae7-4457-a6ac-88e8197d6bc2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.295688 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.297640 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "824eb81c-fae7-4457-a6ac-88e8197d6bc2" (UID: "824eb81c-fae7-4457-a6ac-88e8197d6bc2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.306550 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/824eb81c-fae7-4457-a6ac-88e8197d6bc2-kube-api-access-p4nwj" (OuterVolumeSpecName: "kube-api-access-p4nwj") pod "824eb81c-fae7-4457-a6ac-88e8197d6bc2" (UID: "824eb81c-fae7-4457-a6ac-88e8197d6bc2"). InnerVolumeSpecName "kube-api-access-p4nwj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.312586 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-scripts" (OuterVolumeSpecName: "scripts") pod "824eb81c-fae7-4457-a6ac-88e8197d6bc2" (UID: "824eb81c-fae7-4457-a6ac-88e8197d6bc2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.350412 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "824eb81c-fae7-4457-a6ac-88e8197d6bc2" (UID: "824eb81c-fae7-4457-a6ac-88e8197d6bc2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.389250 4787 generic.go:334] "Generic (PLEG): container finished" podID="c19a0911-5c3e-4c36-96ec-064cd73b8bde" containerID="956351eadf8c23b4abdbbfb3c30e61772f4c35e4ebe96bdb403d10a529f50e2d" exitCode=0 Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.389335 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" event={"ID":"c19a0911-5c3e-4c36-96ec-064cd73b8bde","Type":"ContainerDied","Data":"956351eadf8c23b4abdbbfb3c30e61772f4c35e4ebe96bdb403d10a529f50e2d"} Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.399587 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.399611 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4nwj\" (UniqueName: \"kubernetes.io/projected/824eb81c-fae7-4457-a6ac-88e8197d6bc2-kube-api-access-p4nwj\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.399621 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/824eb81c-fae7-4457-a6ac-88e8197d6bc2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.399629 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.408900 4787 generic.go:334] "Generic (PLEG): container finished" podID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerID="e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6" exitCode=0 Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.409233 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.409319 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"824eb81c-fae7-4457-a6ac-88e8197d6bc2","Type":"ContainerDied","Data":"e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6"} Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.409363 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"824eb81c-fae7-4457-a6ac-88e8197d6bc2","Type":"ContainerDied","Data":"c9d3968f8272a2653916edea12140105d9eaecf96bf52184812426e07f56363f"} Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.409384 4787 scope.go:117] "RemoveContainer" containerID="7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.471975 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "824eb81c-fae7-4457-a6ac-88e8197d6bc2" (UID: "824eb81c-fae7-4457-a6ac-88e8197d6bc2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.476032 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.505454 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.522312 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-config-data" (OuterVolumeSpecName: "config-data") pod "824eb81c-fae7-4457-a6ac-88e8197d6bc2" (UID: "824eb81c-fae7-4457-a6ac-88e8197d6bc2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.607507 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/824eb81c-fae7-4457-a6ac-88e8197d6bc2-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.617299 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bs2bm"] Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.629826 4787 scope.go:117] "RemoveContainer" containerID="6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.649320 4787 scope.go:117] "RemoveContainer" containerID="fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.687921 4787 scope.go:117] "RemoveContainer" containerID="e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.718067 4787 scope.go:117] "RemoveContainer" containerID="7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1" Dec 03 17:39:06 crc kubenswrapper[4787]: E1203 17:39:06.718717 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1\": container with ID starting with 7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1 not found: ID does not exist" containerID="7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.718751 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1"} err="failed to get container status \"7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1\": rpc error: code = NotFound desc = could not find container \"7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1\": container with ID starting with 7b4cdc059a34d569eedb206d75f563d293007fadaee85f46065f6aead2c508f1 not found: ID does not exist" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.718771 4787 scope.go:117] "RemoveContainer" containerID="6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1" Dec 03 17:39:06 crc kubenswrapper[4787]: E1203 17:39:06.719203 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1\": container with ID starting with 6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1 not found: ID does not exist" containerID="6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.719249 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1"} err="failed to get container status \"6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1\": rpc error: code = NotFound desc = could not find container \"6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1\": container with ID starting with 6e59d1022082ddb63edb2bd24078a841a9d946137610a3fddba39e2701521ea1 not found: ID does not exist" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.719280 4787 scope.go:117] "RemoveContainer" containerID="fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e" Dec 03 17:39:06 crc kubenswrapper[4787]: E1203 17:39:06.719588 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e\": container with ID starting with fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e not found: ID does not exist" containerID="fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.719617 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e"} err="failed to get container status \"fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e\": rpc error: code = NotFound desc = could not find container \"fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e\": container with ID starting with fcda5cc03474d54bab95574fa92d3e6d4c8fe808c22340f60873e2c1b92aab1e not found: ID does not exist" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.719635 4787 scope.go:117] "RemoveContainer" containerID="e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6" Dec 03 17:39:06 crc kubenswrapper[4787]: E1203 17:39:06.719905 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6\": container with ID starting with e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6 not found: ID does not exist" containerID="e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.719930 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6"} err="failed to get container status \"e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6\": rpc error: code = NotFound desc = could not find container \"e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6\": container with ID starting with e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6 not found: ID does not exist" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.757992 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.786587 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.807075 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:39:06 crc kubenswrapper[4787]: E1203 17:39:06.807541 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="proxy-httpd" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.807552 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="proxy-httpd" Dec 03 17:39:06 crc kubenswrapper[4787]: E1203 17:39:06.807568 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="sg-core" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.807574 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="sg-core" Dec 03 17:39:06 crc kubenswrapper[4787]: E1203 17:39:06.807590 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="ceilometer-central-agent" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.807598 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="ceilometer-central-agent" Dec 03 17:39:06 crc kubenswrapper[4787]: E1203 17:39:06.807617 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="ceilometer-notification-agent" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.807622 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="ceilometer-notification-agent" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.807802 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="sg-core" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.807819 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="ceilometer-central-agent" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.807841 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="ceilometer-notification-agent" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.807854 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" containerName="proxy-httpd" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.809690 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.813864 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.814236 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.837279 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.859700 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.915939 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-config\") pod \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.915993 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-nb\") pod \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.916078 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-swift-storage-0\") pod \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.916135 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-sb\") pod \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.916228 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcq9p\" (UniqueName: \"kubernetes.io/projected/c19a0911-5c3e-4c36-96ec-064cd73b8bde-kube-api-access-qcq9p\") pod \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.916278 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-svc\") pod \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\" (UID: \"c19a0911-5c3e-4c36-96ec-064cd73b8bde\") " Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.916600 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-log-httpd\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.916677 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dp5vg\" (UniqueName: \"kubernetes.io/projected/f923a880-083f-464a-803a-a4a47662d75f-kube-api-access-dp5vg\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.916700 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.916766 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-run-httpd\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.916860 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.916909 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-config-data\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.916972 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-scripts\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.925240 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c19a0911-5c3e-4c36-96ec-064cd73b8bde-kube-api-access-qcq9p" (OuterVolumeSpecName: "kube-api-access-qcq9p") pod "c19a0911-5c3e-4c36-96ec-064cd73b8bde" (UID: "c19a0911-5c3e-4c36-96ec-064cd73b8bde"). InnerVolumeSpecName "kube-api-access-qcq9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:06 crc kubenswrapper[4787]: I1203 17:39:06.991382 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "c19a0911-5c3e-4c36-96ec-064cd73b8bde" (UID: "c19a0911-5c3e-4c36-96ec-064cd73b8bde"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.005171 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "c19a0911-5c3e-4c36-96ec-064cd73b8bde" (UID: "c19a0911-5c3e-4c36-96ec-064cd73b8bde"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.009857 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "c19a0911-5c3e-4c36-96ec-064cd73b8bde" (UID: "c19a0911-5c3e-4c36-96ec-064cd73b8bde"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.012341 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-config" (OuterVolumeSpecName: "config") pod "c19a0911-5c3e-4c36-96ec-064cd73b8bde" (UID: "c19a0911-5c3e-4c36-96ec-064cd73b8bde"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.015305 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "c19a0911-5c3e-4c36-96ec-064cd73b8bde" (UID: "c19a0911-5c3e-4c36-96ec-064cd73b8bde"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019267 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019325 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-config-data\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019378 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-scripts\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019432 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-log-httpd\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019473 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dp5vg\" (UniqueName: \"kubernetes.io/projected/f923a880-083f-464a-803a-a4a47662d75f-kube-api-access-dp5vg\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019493 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019535 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-run-httpd\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019608 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcq9p\" (UniqueName: \"kubernetes.io/projected/c19a0911-5c3e-4c36-96ec-064cd73b8bde-kube-api-access-qcq9p\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019624 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019636 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019646 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019654 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.019662 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/c19a0911-5c3e-4c36-96ec-064cd73b8bde-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.020092 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-run-httpd\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.020425 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-log-httpd\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.023359 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.023381 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-config-data\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.023764 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.024220 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-scripts\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.036201 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dp5vg\" (UniqueName: \"kubernetes.io/projected/f923a880-083f-464a-803a-a4a47662d75f-kube-api-access-dp5vg\") pod \"ceilometer-0\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.141154 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.425502 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" event={"ID":"c19a0911-5c3e-4c36-96ec-064cd73b8bde","Type":"ContainerDied","Data":"aed22493ce8c642aa3c7d98467024e967fa4209bbc200661d6d2c7dad904540f"} Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.425518 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-9b86998b5-6w5hx" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.425849 4787 scope.go:117] "RemoveContainer" containerID="956351eadf8c23b4abdbbfb3c30e61772f4c35e4ebe96bdb403d10a529f50e2d" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.487869 4787 scope.go:117] "RemoveContainer" containerID="db4c2ec046c25c551a48c23635c66b7b80e0d6a76dad5ada71323359d4355caf" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.501157 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6w5hx"] Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.507726 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-9b86998b5-6w5hx"] Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.651565 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.778865 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="824eb81c-fae7-4457-a6ac-88e8197d6bc2" path="/var/lib/kubelet/pods/824eb81c-fae7-4457-a6ac-88e8197d6bc2/volumes" Dec 03 17:39:07 crc kubenswrapper[4787]: I1203 17:39:07.779911 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c19a0911-5c3e-4c36-96ec-064cd73b8bde" path="/var/lib/kubelet/pods/c19a0911-5c3e-4c36-96ec-064cd73b8bde/volumes" Dec 03 17:39:08 crc kubenswrapper[4787]: I1203 17:39:08.445420 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f923a880-083f-464a-803a-a4a47662d75f","Type":"ContainerStarted","Data":"4393b805a6c73c8c328ce96a80287c5d9d89704a98d6c82405d251950beafa99"} Dec 03 17:39:08 crc kubenswrapper[4787]: I1203 17:39:08.445766 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f923a880-083f-464a-803a-a4a47662d75f","Type":"ContainerStarted","Data":"3d02b33644b8efdd4cfe2a93f12a28e05b749e9947c74f01017e5cd54dfe8d93"} Dec 03 17:39:08 crc kubenswrapper[4787]: I1203 17:39:08.447727 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-bs2bm" podUID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" containerName="registry-server" containerID="cri-o://90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec" gracePeriod=2 Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.076891 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.175720 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-utilities\") pod \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.176156 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4qzsn\" (UniqueName: \"kubernetes.io/projected/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-kube-api-access-4qzsn\") pod \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.176199 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-catalog-content\") pod \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\" (UID: \"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae\") " Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.177694 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-utilities" (OuterVolumeSpecName: "utilities") pod "6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" (UID: "6365945a-9c61-4f6d-a4dc-d1857ff8a1ae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.185098 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-kube-api-access-4qzsn" (OuterVolumeSpecName: "kube-api-access-4qzsn") pod "6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" (UID: "6365945a-9c61-4f6d-a4dc-d1857ff8a1ae"). InnerVolumeSpecName "kube-api-access-4qzsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.259482 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" (UID: "6365945a-9c61-4f6d-a4dc-d1857ff8a1ae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.279047 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.279095 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.279108 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4qzsn\" (UniqueName: \"kubernetes.io/projected/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae-kube-api-access-4qzsn\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.459195 4787 generic.go:334] "Generic (PLEG): container finished" podID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" containerID="90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec" exitCode=0 Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.459221 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs2bm" event={"ID":"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae","Type":"ContainerDied","Data":"90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec"} Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.459254 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-bs2bm" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.459285 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-bs2bm" event={"ID":"6365945a-9c61-4f6d-a4dc-d1857ff8a1ae","Type":"ContainerDied","Data":"8e53dba77da2c4af6ea54e266c59e670d94070c8a9d4c3288028485832dd9bc5"} Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.459303 4787 scope.go:117] "RemoveContainer" containerID="90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.461401 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f923a880-083f-464a-803a-a4a47662d75f","Type":"ContainerStarted","Data":"dbe767f04114b5df04bc3a17a2c3c49e5e55b22cb7e8e16c2b6f320d58df3607"} Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.490340 4787 scope.go:117] "RemoveContainer" containerID="f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.501148 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-bs2bm"] Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.510741 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-bs2bm"] Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.532900 4787 scope.go:117] "RemoveContainer" containerID="230cf3cfeebe79f85db9a716451a99978d9a8fdcb219c52d6ffd18da5310f7e8" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.588132 4787 scope.go:117] "RemoveContainer" containerID="90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec" Dec 03 17:39:09 crc kubenswrapper[4787]: E1203 17:39:09.588803 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec\": container with ID starting with 90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec not found: ID does not exist" containerID="90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.588850 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec"} err="failed to get container status \"90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec\": rpc error: code = NotFound desc = could not find container \"90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec\": container with ID starting with 90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec not found: ID does not exist" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.588883 4787 scope.go:117] "RemoveContainer" containerID="f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399" Dec 03 17:39:09 crc kubenswrapper[4787]: E1203 17:39:09.589196 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399\": container with ID starting with f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399 not found: ID does not exist" containerID="f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.589233 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399"} err="failed to get container status \"f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399\": rpc error: code = NotFound desc = could not find container \"f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399\": container with ID starting with f8a7e5c0c8e647b874d790255928013677fd2a9aefb02c1db43c84179c5d0399 not found: ID does not exist" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.589255 4787 scope.go:117] "RemoveContainer" containerID="230cf3cfeebe79f85db9a716451a99978d9a8fdcb219c52d6ffd18da5310f7e8" Dec 03 17:39:09 crc kubenswrapper[4787]: E1203 17:39:09.590319 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"230cf3cfeebe79f85db9a716451a99978d9a8fdcb219c52d6ffd18da5310f7e8\": container with ID starting with 230cf3cfeebe79f85db9a716451a99978d9a8fdcb219c52d6ffd18da5310f7e8 not found: ID does not exist" containerID="230cf3cfeebe79f85db9a716451a99978d9a8fdcb219c52d6ffd18da5310f7e8" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.590362 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"230cf3cfeebe79f85db9a716451a99978d9a8fdcb219c52d6ffd18da5310f7e8"} err="failed to get container status \"230cf3cfeebe79f85db9a716451a99978d9a8fdcb219c52d6ffd18da5310f7e8\": rpc error: code = NotFound desc = could not find container \"230cf3cfeebe79f85db9a716451a99978d9a8fdcb219c52d6ffd18da5310f7e8\": container with ID starting with 230cf3cfeebe79f85db9a716451a99978d9a8fdcb219c52d6ffd18da5310f7e8 not found: ID does not exist" Dec 03 17:39:09 crc kubenswrapper[4787]: I1203 17:39:09.779345 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" path="/var/lib/kubelet/pods/6365945a-9c61-4f6d-a4dc-d1857ff8a1ae/volumes" Dec 03 17:39:10 crc kubenswrapper[4787]: I1203 17:39:10.482941 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f923a880-083f-464a-803a-a4a47662d75f","Type":"ContainerStarted","Data":"9d438feea66735f410e7fffbd0a1cfb3b45bedb05623a2b1fed636c3a364b6f1"} Dec 03 17:39:11 crc kubenswrapper[4787]: I1203 17:39:11.507963 4787 generic.go:334] "Generic (PLEG): container finished" podID="a28737aa-7ebe-4dbe-967d-51376202958e" containerID="1a00a16219839ac3284d87eacc77080bb5386e413634aaad555e7c56871e39f0" exitCode=0 Dec 03 17:39:11 crc kubenswrapper[4787]: I1203 17:39:11.508228 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7gzqf" event={"ID":"a28737aa-7ebe-4dbe-967d-51376202958e","Type":"ContainerDied","Data":"1a00a16219839ac3284d87eacc77080bb5386e413634aaad555e7c56871e39f0"} Dec 03 17:39:12 crc kubenswrapper[4787]: I1203 17:39:12.521989 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f923a880-083f-464a-803a-a4a47662d75f","Type":"ContainerStarted","Data":"ace71b94e4f14fa0b698938b04697e85cf9cba33cd48f6ffb9680582947c12c3"} Dec 03 17:39:12 crc kubenswrapper[4787]: I1203 17:39:12.522481 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:39:12 crc kubenswrapper[4787]: I1203 17:39:12.555592 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.7345983289999998 podStartE2EDuration="6.555572825s" podCreationTimestamp="2025-12-03 17:39:06 +0000 UTC" firstStartedPulling="2025-12-03 17:39:07.660547697 +0000 UTC m=+1564.478018656" lastFinishedPulling="2025-12-03 17:39:11.481522193 +0000 UTC m=+1568.298993152" observedRunningTime="2025-12-03 17:39:12.545133451 +0000 UTC m=+1569.362604410" watchObservedRunningTime="2025-12-03 17:39:12.555572825 +0000 UTC m=+1569.373043794" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.007121 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.080608 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-scripts\") pod \"a28737aa-7ebe-4dbe-967d-51376202958e\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.080747 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bzx5d\" (UniqueName: \"kubernetes.io/projected/a28737aa-7ebe-4dbe-967d-51376202958e-kube-api-access-bzx5d\") pod \"a28737aa-7ebe-4dbe-967d-51376202958e\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.080802 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-combined-ca-bundle\") pod \"a28737aa-7ebe-4dbe-967d-51376202958e\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.080860 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-config-data\") pod \"a28737aa-7ebe-4dbe-967d-51376202958e\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.086233 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a28737aa-7ebe-4dbe-967d-51376202958e-kube-api-access-bzx5d" (OuterVolumeSpecName: "kube-api-access-bzx5d") pod "a28737aa-7ebe-4dbe-967d-51376202958e" (UID: "a28737aa-7ebe-4dbe-967d-51376202958e"). InnerVolumeSpecName "kube-api-access-bzx5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.099784 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-scripts" (OuterVolumeSpecName: "scripts") pod "a28737aa-7ebe-4dbe-967d-51376202958e" (UID: "a28737aa-7ebe-4dbe-967d-51376202958e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:13 crc kubenswrapper[4787]: E1203 17:39:13.110160 4787 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-combined-ca-bundle podName:a28737aa-7ebe-4dbe-967d-51376202958e nodeName:}" failed. No retries permitted until 2025-12-03 17:39:13.610116805 +0000 UTC m=+1570.427587774 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "combined-ca-bundle" (UniqueName: "kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-combined-ca-bundle") pod "a28737aa-7ebe-4dbe-967d-51376202958e" (UID: "a28737aa-7ebe-4dbe-967d-51376202958e") : error deleting /var/lib/kubelet/pods/a28737aa-7ebe-4dbe-967d-51376202958e/volume-subpaths: remove /var/lib/kubelet/pods/a28737aa-7ebe-4dbe-967d-51376202958e/volume-subpaths: no such file or directory Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.113343 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-config-data" (OuterVolumeSpecName: "config-data") pod "a28737aa-7ebe-4dbe-967d-51376202958e" (UID: "a28737aa-7ebe-4dbe-967d-51376202958e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.184152 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.184732 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.184747 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bzx5d\" (UniqueName: \"kubernetes.io/projected/a28737aa-7ebe-4dbe-967d-51376202958e-kube-api-access-bzx5d\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.539686 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-7gzqf" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.539777 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-7gzqf" event={"ID":"a28737aa-7ebe-4dbe-967d-51376202958e","Type":"ContainerDied","Data":"0f160cc92d9cbc35f7e03821298f905df77500da64d0e2b6ffa5f3ca462bbb79"} Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.539837 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0f160cc92d9cbc35f7e03821298f905df77500da64d0e2b6ffa5f3ca462bbb79" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.695521 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-combined-ca-bundle\") pod \"a28737aa-7ebe-4dbe-967d-51376202958e\" (UID: \"a28737aa-7ebe-4dbe-967d-51376202958e\") " Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.715643 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.715692 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.717281 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "a28737aa-7ebe-4dbe-967d-51376202958e" (UID: "a28737aa-7ebe-4dbe-967d-51376202958e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.782405 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.798834 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a28737aa-7ebe-4dbe-967d-51376202958e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.823506 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.823751 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="c18af891-9fd6-48ea-a881-357223dbabbe" containerName="nova-scheduler-scheduler" containerID="cri-o://3b0937001aa222d83f9c5e71a21aae35441a160c418fc4e451e54d73c16f2796" gracePeriod=30 Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.879869 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.880134 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-log" containerID="cri-o://8b1452abe4a84ea337f4eca5edd118c0aaab837bc5b6d8ae4fe654f7a14ef2b4" gracePeriod=30 Dec 03 17:39:13 crc kubenswrapper[4787]: I1203 17:39:13.880252 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-metadata" containerID="cri-o://8d7b157b2eb22560a8597b21b7b03924b206ff3e05a9c5677d380b3234867a49" gracePeriod=30 Dec 03 17:39:14 crc kubenswrapper[4787]: I1203 17:39:14.564054 4787 generic.go:334] "Generic (PLEG): container finished" podID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerID="8b1452abe4a84ea337f4eca5edd118c0aaab837bc5b6d8ae4fe654f7a14ef2b4" exitCode=143 Dec 03 17:39:14 crc kubenswrapper[4787]: I1203 17:39:14.564140 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"695faa7a-8e43-4277-b206-4c27ae5fd3a5","Type":"ContainerDied","Data":"8b1452abe4a84ea337f4eca5edd118c0aaab837bc5b6d8ae4fe654f7a14ef2b4"} Dec 03 17:39:14 crc kubenswrapper[4787]: I1203 17:39:14.564606 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerName="nova-api-log" containerID="cri-o://6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430" gracePeriod=30 Dec 03 17:39:14 crc kubenswrapper[4787]: I1203 17:39:14.564690 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerName="nova-api-api" containerID="cri-o://0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b" gracePeriod=30 Dec 03 17:39:14 crc kubenswrapper[4787]: I1203 17:39:14.573528 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.241:8774/\": EOF" Dec 03 17:39:14 crc kubenswrapper[4787]: I1203 17:39:14.573553 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.241:8774/\": EOF" Dec 03 17:39:15 crc kubenswrapper[4787]: I1203 17:39:15.640270 4787 generic.go:334] "Generic (PLEG): container finished" podID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerID="6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430" exitCode=143 Dec 03 17:39:15 crc kubenswrapper[4787]: I1203 17:39:15.640330 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e","Type":"ContainerDied","Data":"6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430"} Dec 03 17:39:16 crc kubenswrapper[4787]: E1203 17:39:16.392911 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b0937001aa222d83f9c5e71a21aae35441a160c418fc4e451e54d73c16f2796" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:39:16 crc kubenswrapper[4787]: E1203 17:39:16.396455 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b0937001aa222d83f9c5e71a21aae35441a160c418fc4e451e54d73c16f2796" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:39:16 crc kubenswrapper[4787]: E1203 17:39:16.399042 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="3b0937001aa222d83f9c5e71a21aae35441a160c418fc4e451e54d73c16f2796" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 03 17:39:16 crc kubenswrapper[4787]: E1203 17:39:16.399096 4787 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="c18af891-9fd6-48ea-a881-357223dbabbe" containerName="nova-scheduler-scheduler" Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.029752 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.230:8775/\": read tcp 10.217.0.2:54308->10.217.0.230:8775: read: connection reset by peer" Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.029749 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.230:8775/\": read tcp 10.217.0.2:54310->10.217.0.230:8775: read: connection reset by peer" Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.666089 4787 generic.go:334] "Generic (PLEG): container finished" podID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerID="8d7b157b2eb22560a8597b21b7b03924b206ff3e05a9c5677d380b3234867a49" exitCode=0 Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.666189 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"695faa7a-8e43-4277-b206-4c27ae5fd3a5","Type":"ContainerDied","Data":"8d7b157b2eb22560a8597b21b7b03924b206ff3e05a9c5677d380b3234867a49"} Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.777705 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.947496 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-config-data\") pod \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.947959 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/695faa7a-8e43-4277-b206-4c27ae5fd3a5-logs\") pod \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.948032 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzb9v\" (UniqueName: \"kubernetes.io/projected/695faa7a-8e43-4277-b206-4c27ae5fd3a5-kube-api-access-mzb9v\") pod \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.948057 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-nova-metadata-tls-certs\") pod \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.948117 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-combined-ca-bundle\") pod \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\" (UID: \"695faa7a-8e43-4277-b206-4c27ae5fd3a5\") " Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.951066 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/695faa7a-8e43-4277-b206-4c27ae5fd3a5-logs" (OuterVolumeSpecName: "logs") pod "695faa7a-8e43-4277-b206-4c27ae5fd3a5" (UID: "695faa7a-8e43-4277-b206-4c27ae5fd3a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.957817 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/695faa7a-8e43-4277-b206-4c27ae5fd3a5-kube-api-access-mzb9v" (OuterVolumeSpecName: "kube-api-access-mzb9v") pod "695faa7a-8e43-4277-b206-4c27ae5fd3a5" (UID: "695faa7a-8e43-4277-b206-4c27ae5fd3a5"). InnerVolumeSpecName "kube-api-access-mzb9v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.982576 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-config-data" (OuterVolumeSpecName: "config-data") pod "695faa7a-8e43-4277-b206-4c27ae5fd3a5" (UID: "695faa7a-8e43-4277-b206-4c27ae5fd3a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:17 crc kubenswrapper[4787]: I1203 17:39:17.991616 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "695faa7a-8e43-4277-b206-4c27ae5fd3a5" (UID: "695faa7a-8e43-4277-b206-4c27ae5fd3a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.022950 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "695faa7a-8e43-4277-b206-4c27ae5fd3a5" (UID: "695faa7a-8e43-4277-b206-4c27ae5fd3a5"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.050343 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.050378 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/695faa7a-8e43-4277-b206-4c27ae5fd3a5-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.050391 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzb9v\" (UniqueName: \"kubernetes.io/projected/695faa7a-8e43-4277-b206-4c27ae5fd3a5-kube-api-access-mzb9v\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.050404 4787 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.050415 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/695faa7a-8e43-4277-b206-4c27ae5fd3a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.682782 4787 generic.go:334] "Generic (PLEG): container finished" podID="c18af891-9fd6-48ea-a881-357223dbabbe" containerID="3b0937001aa222d83f9c5e71a21aae35441a160c418fc4e451e54d73c16f2796" exitCode=0 Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.683107 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c18af891-9fd6-48ea-a881-357223dbabbe","Type":"ContainerDied","Data":"3b0937001aa222d83f9c5e71a21aae35441a160c418fc4e451e54d73c16f2796"} Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.685223 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"695faa7a-8e43-4277-b206-4c27ae5fd3a5","Type":"ContainerDied","Data":"a27829b13ab47b4cedd6e0e3cadfe60456e74eb38a13383d361d6c6aee49dcf6"} Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.685254 4787 scope.go:117] "RemoveContainer" containerID="8d7b157b2eb22560a8597b21b7b03924b206ff3e05a9c5677d380b3234867a49" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.685417 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.718546 4787 scope.go:117] "RemoveContainer" containerID="8b1452abe4a84ea337f4eca5edd118c0aaab837bc5b6d8ae4fe654f7a14ef2b4" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.738822 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.808445 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.832098 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:39:18 crc kubenswrapper[4787]: E1203 17:39:18.832627 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19a0911-5c3e-4c36-96ec-064cd73b8bde" containerName="dnsmasq-dns" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.832646 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19a0911-5c3e-4c36-96ec-064cd73b8bde" containerName="dnsmasq-dns" Dec 03 17:39:18 crc kubenswrapper[4787]: E1203 17:39:18.832666 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-log" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.832672 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-log" Dec 03 17:39:18 crc kubenswrapper[4787]: E1203 17:39:18.832690 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c19a0911-5c3e-4c36-96ec-064cd73b8bde" containerName="init" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.832696 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c19a0911-5c3e-4c36-96ec-064cd73b8bde" containerName="init" Dec 03 17:39:18 crc kubenswrapper[4787]: E1203 17:39:18.832705 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" containerName="extract-content" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.832710 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" containerName="extract-content" Dec 03 17:39:18 crc kubenswrapper[4787]: E1203 17:39:18.832717 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-metadata" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.832723 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-metadata" Dec 03 17:39:18 crc kubenswrapper[4787]: E1203 17:39:18.832738 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" containerName="extract-utilities" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.832767 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" containerName="extract-utilities" Dec 03 17:39:18 crc kubenswrapper[4787]: E1203 17:39:18.832784 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a28737aa-7ebe-4dbe-967d-51376202958e" containerName="nova-manage" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.832789 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a28737aa-7ebe-4dbe-967d-51376202958e" containerName="nova-manage" Dec 03 17:39:18 crc kubenswrapper[4787]: E1203 17:39:18.832829 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" containerName="registry-server" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.832834 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" containerName="registry-server" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.833135 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-log" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.833161 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c19a0911-5c3e-4c36-96ec-064cd73b8bde" containerName="dnsmasq-dns" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.833173 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6365945a-9c61-4f6d-a4dc-d1857ff8a1ae" containerName="registry-server" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.833201 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a28737aa-7ebe-4dbe-967d-51376202958e" containerName="nova-manage" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.833214 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" containerName="nova-metadata-metadata" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.834933 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.838494 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.838672 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.844889 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.990114 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be1a822f-d457-417d-9f15-7c3f5b309d7c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.990489 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be1a822f-d457-417d-9f15-7c3f5b309d7c-logs\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.990565 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1a822f-d457-417d-9f15-7c3f5b309d7c-config-data\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.990658 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hbvvv\" (UniqueName: \"kubernetes.io/projected/be1a822f-d457-417d-9f15-7c3f5b309d7c-kube-api-access-hbvvv\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:18 crc kubenswrapper[4787]: I1203 17:39:18.990696 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1a822f-d457-417d-9f15-7c3f5b309d7c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.093569 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hbvvv\" (UniqueName: \"kubernetes.io/projected/be1a822f-d457-417d-9f15-7c3f5b309d7c-kube-api-access-hbvvv\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.093627 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1a822f-d457-417d-9f15-7c3f5b309d7c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.093771 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be1a822f-d457-417d-9f15-7c3f5b309d7c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.093843 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be1a822f-d457-417d-9f15-7c3f5b309d7c-logs\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.093977 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1a822f-d457-417d-9f15-7c3f5b309d7c-config-data\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.094392 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be1a822f-d457-417d-9f15-7c3f5b309d7c-logs\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.099308 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/be1a822f-d457-417d-9f15-7c3f5b309d7c-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.099810 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be1a822f-d457-417d-9f15-7c3f5b309d7c-config-data\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.100260 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be1a822f-d457-417d-9f15-7c3f5b309d7c-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.118672 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hbvvv\" (UniqueName: \"kubernetes.io/projected/be1a822f-d457-417d-9f15-7c3f5b309d7c-kube-api-access-hbvvv\") pod \"nova-metadata-0\" (UID: \"be1a822f-d457-417d-9f15-7c3f5b309d7c\") " pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.163676 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.266605 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.397424 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ptzwb\" (UniqueName: \"kubernetes.io/projected/c18af891-9fd6-48ea-a881-357223dbabbe-kube-api-access-ptzwb\") pod \"c18af891-9fd6-48ea-a881-357223dbabbe\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.397825 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-config-data\") pod \"c18af891-9fd6-48ea-a881-357223dbabbe\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.397948 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-combined-ca-bundle\") pod \"c18af891-9fd6-48ea-a881-357223dbabbe\" (UID: \"c18af891-9fd6-48ea-a881-357223dbabbe\") " Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.401895 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c18af891-9fd6-48ea-a881-357223dbabbe-kube-api-access-ptzwb" (OuterVolumeSpecName: "kube-api-access-ptzwb") pod "c18af891-9fd6-48ea-a881-357223dbabbe" (UID: "c18af891-9fd6-48ea-a881-357223dbabbe"). InnerVolumeSpecName "kube-api-access-ptzwb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.429496 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-config-data" (OuterVolumeSpecName: "config-data") pod "c18af891-9fd6-48ea-a881-357223dbabbe" (UID: "c18af891-9fd6-48ea-a881-357223dbabbe"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.445898 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c18af891-9fd6-48ea-a881-357223dbabbe" (UID: "c18af891-9fd6-48ea-a881-357223dbabbe"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.500555 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.500583 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ptzwb\" (UniqueName: \"kubernetes.io/projected/c18af891-9fd6-48ea-a881-357223dbabbe-kube-api-access-ptzwb\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.500592 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c18af891-9fd6-48ea-a881-357223dbabbe-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.625100 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.709907 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"c18af891-9fd6-48ea-a881-357223dbabbe","Type":"ContainerDied","Data":"53e508d9d1d7fe81a8de542fb46d7c9c99e3491eb4e463ad852189f0191a86ae"} Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.709969 4787 scope.go:117] "RemoveContainer" containerID="3b0937001aa222d83f9c5e71a21aae35441a160c418fc4e451e54d73c16f2796" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.710106 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.732526 4787 generic.go:334] "Generic (PLEG): container finished" podID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerID="7ea84ed69f546d0057496b37bad2ff9e4768d95f9ba2190731bd6e284ffa7a63" exitCode=137 Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.732555 4787 generic.go:334] "Generic (PLEG): container finished" podID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerID="620f3781c6005f73c56e92517da3a7998413f902ca506cca186ebe2fb3525db8" exitCode=137 Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.732597 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4282e0c2-7846-4f1f-afeb-076726cfe29d","Type":"ContainerDied","Data":"7ea84ed69f546d0057496b37bad2ff9e4768d95f9ba2190731bd6e284ffa7a63"} Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.732625 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4282e0c2-7846-4f1f-afeb-076726cfe29d","Type":"ContainerDied","Data":"620f3781c6005f73c56e92517da3a7998413f902ca506cca186ebe2fb3525db8"} Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.750821 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be1a822f-d457-417d-9f15-7c3f5b309d7c","Type":"ContainerStarted","Data":"09abb1ec0bbf218e30c789448eab2308e3dd043715f8c56db0d771e33eac7ab8"} Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.786081 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="695faa7a-8e43-4277-b206-4c27ae5fd3a5" path="/var/lib/kubelet/pods/695faa7a-8e43-4277-b206-4c27ae5fd3a5/volumes" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.793834 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.806091 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.814395 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:39:19 crc kubenswrapper[4787]: E1203 17:39:19.814809 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c18af891-9fd6-48ea-a881-357223dbabbe" containerName="nova-scheduler-scheduler" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.814825 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c18af891-9fd6-48ea-a881-357223dbabbe" containerName="nova-scheduler-scheduler" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.815078 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c18af891-9fd6-48ea-a881-357223dbabbe" containerName="nova-scheduler-scheduler" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.815833 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.819484 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.833781 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.840911 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:39:19 crc kubenswrapper[4787]: E1203 17:39:19.874402 4787 fsHandler.go:119] failed to collect filesystem stats - rootDiskErr: could not stat "/var/lib/containers/storage/overlay/764020f8fe5c7bbb74a304b6629e1fe4686ac26986f25ff584a0ac4d4f2f5e81/diff" to get inode usage: stat /var/lib/containers/storage/overlay/764020f8fe5c7bbb74a304b6629e1fe4686ac26986f25ff584a0ac4d4f2f5e81/diff: no such file or directory, extraDiskErr: could not stat "/var/log/pods/openstack_dnsmasq-dns-9b86998b5-6w5hx_c19a0911-5c3e-4c36-96ec-064cd73b8bde/dnsmasq-dns/0.log" to get inode usage: stat /var/log/pods/openstack_dnsmasq-dns-9b86998b5-6w5hx_c19a0911-5c3e-4c36-96ec-064cd73b8bde/dnsmasq-dns/0.log: no such file or directory Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.916495 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8g279\" (UniqueName: \"kubernetes.io/projected/1a00b714-bec1-411f-a649-c9825253f05e-kube-api-access-8g279\") pod \"nova-scheduler-0\" (UID: \"1a00b714-bec1-411f-a649-c9825253f05e\") " pod="openstack/nova-scheduler-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.916723 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a00b714-bec1-411f-a649-c9825253f05e-config-data\") pod \"nova-scheduler-0\" (UID: \"1a00b714-bec1-411f-a649-c9825253f05e\") " pod="openstack/nova-scheduler-0" Dec 03 17:39:19 crc kubenswrapper[4787]: I1203 17:39:19.916973 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a00b714-bec1-411f-a649-c9825253f05e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a00b714-bec1-411f-a649-c9825253f05e\") " pod="openstack/nova-scheduler-0" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.018956 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4n7j8\" (UniqueName: \"kubernetes.io/projected/4282e0c2-7846-4f1f-afeb-076726cfe29d-kube-api-access-4n7j8\") pod \"4282e0c2-7846-4f1f-afeb-076726cfe29d\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.019047 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-scripts\") pod \"4282e0c2-7846-4f1f-afeb-076726cfe29d\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.019107 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-config-data\") pod \"4282e0c2-7846-4f1f-afeb-076726cfe29d\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.019137 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-combined-ca-bundle\") pod \"4282e0c2-7846-4f1f-afeb-076726cfe29d\" (UID: \"4282e0c2-7846-4f1f-afeb-076726cfe29d\") " Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.019613 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8g279\" (UniqueName: \"kubernetes.io/projected/1a00b714-bec1-411f-a649-c9825253f05e-kube-api-access-8g279\") pod \"nova-scheduler-0\" (UID: \"1a00b714-bec1-411f-a649-c9825253f05e\") " pod="openstack/nova-scheduler-0" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.019687 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a00b714-bec1-411f-a649-c9825253f05e-config-data\") pod \"nova-scheduler-0\" (UID: \"1a00b714-bec1-411f-a649-c9825253f05e\") " pod="openstack/nova-scheduler-0" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.019750 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a00b714-bec1-411f-a649-c9825253f05e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a00b714-bec1-411f-a649-c9825253f05e\") " pod="openstack/nova-scheduler-0" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.023550 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4282e0c2-7846-4f1f-afeb-076726cfe29d-kube-api-access-4n7j8" (OuterVolumeSpecName: "kube-api-access-4n7j8") pod "4282e0c2-7846-4f1f-afeb-076726cfe29d" (UID: "4282e0c2-7846-4f1f-afeb-076726cfe29d"). InnerVolumeSpecName "kube-api-access-4n7j8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.025120 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1a00b714-bec1-411f-a649-c9825253f05e-config-data\") pod \"nova-scheduler-0\" (UID: \"1a00b714-bec1-411f-a649-c9825253f05e\") " pod="openstack/nova-scheduler-0" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.027910 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1a00b714-bec1-411f-a649-c9825253f05e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"1a00b714-bec1-411f-a649-c9825253f05e\") " pod="openstack/nova-scheduler-0" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.036401 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-scripts" (OuterVolumeSpecName: "scripts") pod "4282e0c2-7846-4f1f-afeb-076726cfe29d" (UID: "4282e0c2-7846-4f1f-afeb-076726cfe29d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.040962 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8g279\" (UniqueName: \"kubernetes.io/projected/1a00b714-bec1-411f-a649-c9825253f05e-kube-api-access-8g279\") pod \"nova-scheduler-0\" (UID: \"1a00b714-bec1-411f-a649-c9825253f05e\") " pod="openstack/nova-scheduler-0" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.125420 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4n7j8\" (UniqueName: \"kubernetes.io/projected/4282e0c2-7846-4f1f-afeb-076726cfe29d-kube-api-access-4n7j8\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.126131 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.147752 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.190215 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-config-data" (OuterVolumeSpecName: "config-data") pod "4282e0c2-7846-4f1f-afeb-076726cfe29d" (UID: "4282e0c2-7846-4f1f-afeb-076726cfe29d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.213077 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4282e0c2-7846-4f1f-afeb-076726cfe29d" (UID: "4282e0c2-7846-4f1f-afeb-076726cfe29d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.233851 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.233890 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4282e0c2-7846-4f1f-afeb-076726cfe29d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:20 crc kubenswrapper[4787]: E1203 17:39:20.392593 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod695faa7a_8e43_4277_b206_4c27ae5fd3a5.slice/crio-a27829b13ab47b4cedd6e0e3cadfe60456e74eb38a13383d361d6c6aee49dcf6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc18af891_9fd6_48ea_a881_357223dbabbe.slice/crio-3b0937001aa222d83f9c5e71a21aae35441a160c418fc4e451e54d73c16f2796.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4282e0c2_7846_4f1f_afeb_076726cfe29d.slice/crio-conmon-7ea84ed69f546d0057496b37bad2ff9e4768d95f9ba2190731bd6e284ffa7a63.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6365945a_9c61_4f6d_a4dc_d1857ff8a1ae.slice/crio-conmon-90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4282e0c2_7846_4f1f_afeb_076726cfe29d.slice/crio-7ea84ed69f546d0057496b37bad2ff9e4768d95f9ba2190731bd6e284ffa7a63.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc19a0911_5c3e_4c36_96ec_064cd73b8bde.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod824eb81c_fae7_4457_a6ac_88e8197d6bc2.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc19a0911_5c3e_4c36_96ec_064cd73b8bde.slice/crio-956351eadf8c23b4abdbbfb3c30e61772f4c35e4ebe96bdb403d10a529f50e2d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda28737aa_7ebe_4dbe_967d_51376202958e.slice/crio-conmon-1a00a16219839ac3284d87eacc77080bb5386e413634aaad555e7c56871e39f0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod824eb81c_fae7_4457_a6ac_88e8197d6bc2.slice/crio-conmon-e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod695faa7a_8e43_4277_b206_4c27ae5fd3a5.slice/crio-8b1452abe4a84ea337f4eca5edd118c0aaab837bc5b6d8ae4fe654f7a14ef2b4.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc18af891_9fd6_48ea_a881_357223dbabbe.slice/crio-53e508d9d1d7fe81a8de542fb46d7c9c99e3491eb4e463ad852189f0191a86ae\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc18af891_9fd6_48ea_a881_357223dbabbe.slice/crio-conmon-3b0937001aa222d83f9c5e71a21aae35441a160c418fc4e451e54d73c16f2796.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5ec4bdb_18f2_490f_829b_ccc6d6e1470e.slice/crio-conmon-6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6365945a_9c61_4f6d_a4dc_d1857ff8a1ae.slice/crio-90cd19ece07b8322e474dd0a845198df0ca6b40960ff52a215c3b2e3490aeeec.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4282e0c2_7846_4f1f_afeb_076726cfe29d.slice/crio-620f3781c6005f73c56e92517da3a7998413f902ca506cca186ebe2fb3525db8.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6365945a_9c61_4f6d_a4dc_d1857ff8a1ae.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda28737aa_7ebe_4dbe_967d_51376202958e.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc18af891_9fd6_48ea_a881_357223dbabbe.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6365945a_9c61_4f6d_a4dc_d1857ff8a1ae.slice/crio-8e53dba77da2c4af6ea54e266c59e670d94070c8a9d4c3288028485832dd9bc5\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc19a0911_5c3e_4c36_96ec_064cd73b8bde.slice/crio-conmon-956351eadf8c23b4abdbbfb3c30e61772f4c35e4ebe96bdb403d10a529f50e2d.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod824eb81c_fae7_4457_a6ac_88e8197d6bc2.slice/crio-e660f10ae25671411365dde2027b6807fca180fc9c7a3169cf0c6321b3b606b6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode5ec4bdb_18f2_490f_829b_ccc6d6e1470e.slice/crio-6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod695faa7a_8e43_4277_b206_4c27ae5fd3a5.slice/crio-conmon-8d7b157b2eb22560a8597b21b7b03924b206ff3e05a9c5677d380b3234867a49.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda28737aa_7ebe_4dbe_967d_51376202958e.slice/crio-1a00a16219839ac3284d87eacc77080bb5386e413634aaad555e7c56871e39f0.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda28737aa_7ebe_4dbe_967d_51376202958e.slice/crio-0f160cc92d9cbc35f7e03821298f905df77500da64d0e2b6ffa5f3ca462bbb79\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod824eb81c_fae7_4457_a6ac_88e8197d6bc2.slice/crio-c9d3968f8272a2653916edea12140105d9eaecf96bf52184812426e07f56363f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc19a0911_5c3e_4c36_96ec_064cd73b8bde.slice/crio-aed22493ce8c642aa3c7d98467024e967fa4209bbc200661d6d2c7dad904540f\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod695faa7a_8e43_4277_b206_4c27ae5fd3a5.slice/crio-conmon-8b1452abe4a84ea337f4eca5edd118c0aaab837bc5b6d8ae4fe654f7a14ef2b4.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.683380 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.856997 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-config-data\") pod \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.859320 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-internal-tls-certs\") pod \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.859439 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-29wzw\" (UniqueName: \"kubernetes.io/projected/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-kube-api-access-29wzw\") pod \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.859491 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-public-tls-certs\") pod \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.859554 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-combined-ca-bundle\") pod \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.859631 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-logs\") pod \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\" (UID: \"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e\") " Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.865430 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-logs" (OuterVolumeSpecName: "logs") pod "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" (UID: "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.880689 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.899696 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-kube-api-access-29wzw" (OuterVolumeSpecName: "kube-api-access-29wzw") pod "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" (UID: "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e"). InnerVolumeSpecName "kube-api-access-29wzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.902108 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"4282e0c2-7846-4f1f-afeb-076726cfe29d","Type":"ContainerDied","Data":"a6093dcf7ba99c69f3545111f2d8b1f0472a6d621770fb905f7a7891af96b1ff"} Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.902167 4787 scope.go:117] "RemoveContainer" containerID="7ea84ed69f546d0057496b37bad2ff9e4768d95f9ba2190731bd6e284ffa7a63" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.902371 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.913935 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be1a822f-d457-417d-9f15-7c3f5b309d7c","Type":"ContainerStarted","Data":"adadb99c9114da2bae3336407d80a20fb26b6f0f3a7c30bb3e76143f10f44b26"} Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.913976 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"be1a822f-d457-417d-9f15-7c3f5b309d7c","Type":"ContainerStarted","Data":"907b992f34db5b179f5da120819db7fc27362b874f8d6150af77d6404bf7cca4"} Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.917494 4787 generic.go:334] "Generic (PLEG): container finished" podID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerID="0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b" exitCode=0 Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.917549 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e","Type":"ContainerDied","Data":"0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b"} Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.917571 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"e5ec4bdb-18f2-490f-829b-ccc6d6e1470e","Type":"ContainerDied","Data":"3de0b9dd110f11333a6f2378b94970c8e1660915e9e777f8bb4d38f2f7863fc6"} Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.917807 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.937390 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-config-data" (OuterVolumeSpecName: "config-data") pod "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" (UID: "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.955347 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" (UID: "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.958356 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.958327196 podStartE2EDuration="2.958327196s" podCreationTimestamp="2025-12-03 17:39:18 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:39:20.937808399 +0000 UTC m=+1577.755279358" watchObservedRunningTime="2025-12-03 17:39:20.958327196 +0000 UTC m=+1577.775798155" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.963159 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.963206 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-29wzw\" (UniqueName: \"kubernetes.io/projected/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-kube-api-access-29wzw\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.963222 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.963234 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-logs\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.988665 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" (UID: "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.991410 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 17:39:20 crc kubenswrapper[4787]: I1203 17:39:20.993739 4787 scope.go:117] "RemoveContainer" containerID="620f3781c6005f73c56e92517da3a7998413f902ca506cca186ebe2fb3525db8" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.017319 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.037927 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.039930 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" (UID: "e5ec4bdb-18f2-490f-829b-ccc6d6e1470e"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:21 crc kubenswrapper[4787]: E1203 17:39:21.040335 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-notifier" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040374 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-notifier" Dec 03 17:39:21 crc kubenswrapper[4787]: E1203 17:39:21.040422 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerName="nova-api-log" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040431 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerName="nova-api-log" Dec 03 17:39:21 crc kubenswrapper[4787]: E1203 17:39:21.040445 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-api" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040453 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-api" Dec 03 17:39:21 crc kubenswrapper[4787]: E1203 17:39:21.040487 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-listener" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040495 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-listener" Dec 03 17:39:21 crc kubenswrapper[4787]: E1203 17:39:21.040512 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerName="nova-api-api" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040522 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerName="nova-api-api" Dec 03 17:39:21 crc kubenswrapper[4787]: E1203 17:39:21.040541 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-evaluator" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040572 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-evaluator" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040905 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-listener" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040935 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerName="nova-api-api" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040952 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" containerName="nova-api-log" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040970 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-api" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040986 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-notifier" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.040998 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" containerName="aodh-evaluator" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.044562 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.047062 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.047367 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.047467 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-q4n2s" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.047495 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.047647 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.067109 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.067122 4787 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.067206 4787 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.078089 4787 scope.go:117] "RemoveContainer" containerID="1adb2a520fce01177ef81401295860b153f5aab544859ed528f5654fa3326feb" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.122331 4787 scope.go:117] "RemoveContainer" containerID="fc7909877fcab48f36abd4f3e735bc3e386dede90fc29fb0260a7deda5ac65b4" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.168397 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-scripts\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.168478 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-public-tls-certs\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.168513 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-config-data\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.168544 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phjxc\" (UniqueName: \"kubernetes.io/projected/b51192bf-fe51-4594-ae63-eba136c23535-kube-api-access-phjxc\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.168570 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-combined-ca-bundle\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.168667 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-internal-tls-certs\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.188407 4787 scope.go:117] "RemoveContainer" containerID="0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.270793 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-internal-tls-certs\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.270874 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-scripts\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.270944 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-public-tls-certs\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.270989 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-config-data\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.271044 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phjxc\" (UniqueName: \"kubernetes.io/projected/b51192bf-fe51-4594-ae63-eba136c23535-kube-api-access-phjxc\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.271079 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-combined-ca-bundle\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.273998 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.275691 4787 scope.go:117] "RemoveContainer" containerID="6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.279928 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-public-tls-certs\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.286684 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-scripts\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.286862 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-combined-ca-bundle\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.287559 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-internal-tls-certs\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.293220 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-config-data\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.307593 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phjxc\" (UniqueName: \"kubernetes.io/projected/b51192bf-fe51-4594-ae63-eba136c23535-kube-api-access-phjxc\") pod \"aodh-0\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.316439 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.338287 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.342995 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.349847 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.350907 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.351168 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.351343 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.371729 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.437762 4787 scope.go:117] "RemoveContainer" containerID="0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b" Dec 03 17:39:21 crc kubenswrapper[4787]: E1203 17:39:21.438288 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b\": container with ID starting with 0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b not found: ID does not exist" containerID="0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.438524 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b"} err="failed to get container status \"0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b\": rpc error: code = NotFound desc = could not find container \"0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b\": container with ID starting with 0eae0bcf0e5032ce79ee24526f803609363248ce14a21120c4ec5296f565be9b not found: ID does not exist" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.438550 4787 scope.go:117] "RemoveContainer" containerID="6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430" Dec 03 17:39:21 crc kubenswrapper[4787]: E1203 17:39:21.438920 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430\": container with ID starting with 6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430 not found: ID does not exist" containerID="6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.438949 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430"} err="failed to get container status \"6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430\": rpc error: code = NotFound desc = could not find container \"6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430\": container with ID starting with 6ba9105175bcd36bae3d0c842658e1d4ed0e43d4b1482132f20ab46e9c133430 not found: ID does not exist" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.475962 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-logs\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.476119 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-config-data\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.476182 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nfhl\" (UniqueName: \"kubernetes.io/projected/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-kube-api-access-2nfhl\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.476209 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-internal-tls-certs\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.476241 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.476356 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-public-tls-certs\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.582864 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-public-tls-certs\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.583262 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-logs\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.583308 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-config-data\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.583357 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nfhl\" (UniqueName: \"kubernetes.io/projected/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-kube-api-access-2nfhl\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.583380 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-internal-tls-certs\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.583404 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.583904 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-logs\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.587996 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-config-data\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.596309 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-public-tls-certs\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.597252 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-internal-tls-certs\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.598747 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.599066 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nfhl\" (UniqueName: \"kubernetes.io/projected/98746b52-f9a2-4d4f-ab95-2a1c5d7429ec-kube-api-access-2nfhl\") pod \"nova-api-0\" (UID: \"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec\") " pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.695029 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.783473 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4282e0c2-7846-4f1f-afeb-076726cfe29d" path="/var/lib/kubelet/pods/4282e0c2-7846-4f1f-afeb-076726cfe29d/volumes" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.784844 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c18af891-9fd6-48ea-a881-357223dbabbe" path="/var/lib/kubelet/pods/c18af891-9fd6-48ea-a881-357223dbabbe/volumes" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.785417 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5ec4bdb-18f2-490f-829b-ccc6d6e1470e" path="/var/lib/kubelet/pods/e5ec4bdb-18f2-490f-829b-ccc6d6e1470e/volumes" Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.896112 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.934719 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a00b714-bec1-411f-a649-c9825253f05e","Type":"ContainerStarted","Data":"e4c04acc7da6016ab4966126cddbb8b8be1f41ff5c44b3d1b16a6db0b13bc14d"} Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.934765 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"1a00b714-bec1-411f-a649-c9825253f05e","Type":"ContainerStarted","Data":"c5818385a0f60840493db1848815f38fdf7df5559121f1ef25f35ae7107b233a"} Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.937461 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b51192bf-fe51-4594-ae63-eba136c23535","Type":"ContainerStarted","Data":"70464ed43c1dfd6ac998f0aa03a617d503a935134442b9b4eb750762b885bcb9"} Dec 03 17:39:21 crc kubenswrapper[4787]: I1203 17:39:21.959767 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.959747886 podStartE2EDuration="2.959747886s" podCreationTimestamp="2025-12-03 17:39:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:39:21.957361212 +0000 UTC m=+1578.774832181" watchObservedRunningTime="2025-12-03 17:39:21.959747886 +0000 UTC m=+1578.777218835" Dec 03 17:39:22 crc kubenswrapper[4787]: I1203 17:39:22.168424 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 03 17:39:22 crc kubenswrapper[4787]: I1203 17:39:22.977186 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b51192bf-fe51-4594-ae63-eba136c23535","Type":"ContainerStarted","Data":"58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02"} Dec 03 17:39:22 crc kubenswrapper[4787]: I1203 17:39:22.983376 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec","Type":"ContainerStarted","Data":"135887e038834fa566bf17eda218adca0cf1e686a34dcbc5c6f9318f8bfea3fa"} Dec 03 17:39:22 crc kubenswrapper[4787]: I1203 17:39:22.985260 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec","Type":"ContainerStarted","Data":"947c6872395f4d766c2f0656273681903f4d8abfaa02326ba8041ddbd3685cfb"} Dec 03 17:39:22 crc kubenswrapper[4787]: I1203 17:39:22.985565 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"98746b52-f9a2-4d4f-ab95-2a1c5d7429ec","Type":"ContainerStarted","Data":"185e6f2d1745d3e55fbb4e107fc39299146a10d734e80fe82851f3eda0bebb2c"} Dec 03 17:39:23 crc kubenswrapper[4787]: I1203 17:39:23.010294 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.010271779 podStartE2EDuration="2.010271779s" podCreationTimestamp="2025-12-03 17:39:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:39:23.00661574 +0000 UTC m=+1579.824086699" watchObservedRunningTime="2025-12-03 17:39:23.010271779 +0000 UTC m=+1579.827742738" Dec 03 17:39:23 crc kubenswrapper[4787]: I1203 17:39:23.994469 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b51192bf-fe51-4594-ae63-eba136c23535","Type":"ContainerStarted","Data":"2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67"} Dec 03 17:39:23 crc kubenswrapper[4787]: I1203 17:39:23.995139 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b51192bf-fe51-4594-ae63-eba136c23535","Type":"ContainerStarted","Data":"133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5"} Dec 03 17:39:24 crc kubenswrapper[4787]: I1203 17:39:24.163802 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:39:24 crc kubenswrapper[4787]: I1203 17:39:24.164572 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 03 17:39:25 crc kubenswrapper[4787]: I1203 17:39:25.008308 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b51192bf-fe51-4594-ae63-eba136c23535","Type":"ContainerStarted","Data":"c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea"} Dec 03 17:39:25 crc kubenswrapper[4787]: I1203 17:39:25.037657 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.818778111 podStartE2EDuration="5.037633544s" podCreationTimestamp="2025-12-03 17:39:20 +0000 UTC" firstStartedPulling="2025-12-03 17:39:21.90278423 +0000 UTC m=+1578.720255189" lastFinishedPulling="2025-12-03 17:39:24.121639653 +0000 UTC m=+1580.939110622" observedRunningTime="2025-12-03 17:39:25.028415504 +0000 UTC m=+1581.845886453" watchObservedRunningTime="2025-12-03 17:39:25.037633544 +0000 UTC m=+1581.855104503" Dec 03 17:39:25 crc kubenswrapper[4787]: I1203 17:39:25.148912 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 03 17:39:29 crc kubenswrapper[4787]: I1203 17:39:29.164509 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 17:39:29 crc kubenswrapper[4787]: I1203 17:39:29.165332 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 03 17:39:30 crc kubenswrapper[4787]: I1203 17:39:30.149591 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 03 17:39:30 crc kubenswrapper[4787]: I1203 17:39:30.179197 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="be1a822f-d457-417d-9f15-7c3f5b309d7c" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.244:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:39:30 crc kubenswrapper[4787]: I1203 17:39:30.179239 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="be1a822f-d457-417d-9f15-7c3f5b309d7c" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.244:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:39:30 crc kubenswrapper[4787]: I1203 17:39:30.181227 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 03 17:39:31 crc kubenswrapper[4787]: I1203 17:39:31.105242 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 03 17:39:31 crc kubenswrapper[4787]: I1203 17:39:31.696035 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:39:31 crc kubenswrapper[4787]: I1203 17:39:31.696357 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 03 17:39:32 crc kubenswrapper[4787]: I1203 17:39:32.708250 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="98746b52-f9a2-4d4f-ab95-2a1c5d7429ec" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.247:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:39:32 crc kubenswrapper[4787]: I1203 17:39:32.708235 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="98746b52-f9a2-4d4f-ab95-2a1c5d7429ec" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.247:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 03 17:39:37 crc kubenswrapper[4787]: I1203 17:39:37.150873 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 17:39:39 crc kubenswrapper[4787]: I1203 17:39:39.173070 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 17:39:39 crc kubenswrapper[4787]: I1203 17:39:39.173768 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 03 17:39:39 crc kubenswrapper[4787]: I1203 17:39:39.179294 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 17:39:39 crc kubenswrapper[4787]: I1203 17:39:39.180803 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 03 17:39:41 crc kubenswrapper[4787]: I1203 17:39:41.704579 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 17:39:41 crc kubenswrapper[4787]: I1203 17:39:41.705469 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 17:39:41 crc kubenswrapper[4787]: I1203 17:39:41.709802 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 03 17:39:41 crc kubenswrapper[4787]: I1203 17:39:41.719833 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 17:39:42 crc kubenswrapper[4787]: I1203 17:39:42.193302 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 03 17:39:42 crc kubenswrapper[4787]: I1203 17:39:42.199609 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 03 17:39:42 crc kubenswrapper[4787]: I1203 17:39:42.594935 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:39:42 crc kubenswrapper[4787]: I1203 17:39:42.595174 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="f5a1dad4-bf28-4c9a-987f-c66658f51ae7" containerName="kube-state-metrics" containerID="cri-o://00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8" gracePeriod=30 Dec 03 17:39:42 crc kubenswrapper[4787]: I1203 17:39:42.675000 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 17:39:42 crc kubenswrapper[4787]: I1203 17:39:42.675541 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/mysqld-exporter-0" podUID="561fe2dd-e048-4657-8c0c-2434fe09f35e" containerName="mysqld-exporter" containerID="cri-o://b4de70be6d3d549766d0ae5211c21fd50c58ada0b50b6cdf1f975662db978161" gracePeriod=30 Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.190173 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.215694 4787 generic.go:334] "Generic (PLEG): container finished" podID="561fe2dd-e048-4657-8c0c-2434fe09f35e" containerID="b4de70be6d3d549766d0ae5211c21fd50c58ada0b50b6cdf1f975662db978161" exitCode=2 Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.215753 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"561fe2dd-e048-4657-8c0c-2434fe09f35e","Type":"ContainerDied","Data":"b4de70be6d3d549766d0ae5211c21fd50c58ada0b50b6cdf1f975662db978161"} Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.218639 4787 generic.go:334] "Generic (PLEG): container finished" podID="f5a1dad4-bf28-4c9a-987f-c66658f51ae7" containerID="00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8" exitCode=2 Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.219303 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.219547 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f5a1dad4-bf28-4c9a-987f-c66658f51ae7","Type":"ContainerDied","Data":"00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8"} Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.219571 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f5a1dad4-bf28-4c9a-987f-c66658f51ae7","Type":"ContainerDied","Data":"7f55d21b4081d87824f14b26d6a717db9b12671e2b608b8e71255341c724f08e"} Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.219588 4787 scope.go:117] "RemoveContainer" containerID="00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.268705 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t6977\" (UniqueName: \"kubernetes.io/projected/f5a1dad4-bf28-4c9a-987f-c66658f51ae7-kube-api-access-t6977\") pod \"f5a1dad4-bf28-4c9a-987f-c66658f51ae7\" (UID: \"f5a1dad4-bf28-4c9a-987f-c66658f51ae7\") " Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.279372 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.280267 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5a1dad4-bf28-4c9a-987f-c66658f51ae7-kube-api-access-t6977" (OuterVolumeSpecName: "kube-api-access-t6977") pod "f5a1dad4-bf28-4c9a-987f-c66658f51ae7" (UID: "f5a1dad4-bf28-4c9a-987f-c66658f51ae7"). InnerVolumeSpecName "kube-api-access-t6977". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.288155 4787 scope.go:117] "RemoveContainer" containerID="00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8" Dec 03 17:39:43 crc kubenswrapper[4787]: E1203 17:39:43.289900 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8\": container with ID starting with 00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8 not found: ID does not exist" containerID="00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.289938 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8"} err="failed to get container status \"00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8\": rpc error: code = NotFound desc = could not find container \"00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8\": container with ID starting with 00444b30d2300709a4b35064c7c6dd953eb62a5e223fc691fabe3d61ee9cd1e8 not found: ID does not exist" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.370562 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-combined-ca-bundle\") pod \"561fe2dd-e048-4657-8c0c-2434fe09f35e\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.370684 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrf2l\" (UniqueName: \"kubernetes.io/projected/561fe2dd-e048-4657-8c0c-2434fe09f35e-kube-api-access-xrf2l\") pod \"561fe2dd-e048-4657-8c0c-2434fe09f35e\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.370724 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-config-data\") pod \"561fe2dd-e048-4657-8c0c-2434fe09f35e\" (UID: \"561fe2dd-e048-4657-8c0c-2434fe09f35e\") " Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.371301 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t6977\" (UniqueName: \"kubernetes.io/projected/f5a1dad4-bf28-4c9a-987f-c66658f51ae7-kube-api-access-t6977\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.374509 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/561fe2dd-e048-4657-8c0c-2434fe09f35e-kube-api-access-xrf2l" (OuterVolumeSpecName: "kube-api-access-xrf2l") pod "561fe2dd-e048-4657-8c0c-2434fe09f35e" (UID: "561fe2dd-e048-4657-8c0c-2434fe09f35e"). InnerVolumeSpecName "kube-api-access-xrf2l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.404232 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "561fe2dd-e048-4657-8c0c-2434fe09f35e" (UID: "561fe2dd-e048-4657-8c0c-2434fe09f35e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.451771 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-config-data" (OuterVolumeSpecName: "config-data") pod "561fe2dd-e048-4657-8c0c-2434fe09f35e" (UID: "561fe2dd-e048-4657-8c0c-2434fe09f35e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.473507 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrf2l\" (UniqueName: \"kubernetes.io/projected/561fe2dd-e048-4657-8c0c-2434fe09f35e-kube-api-access-xrf2l\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.473753 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.473892 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/561fe2dd-e048-4657-8c0c-2434fe09f35e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.553535 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.563698 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.578318 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:39:43 crc kubenswrapper[4787]: E1203 17:39:43.579134 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5a1dad4-bf28-4c9a-987f-c66658f51ae7" containerName="kube-state-metrics" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.579229 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5a1dad4-bf28-4c9a-987f-c66658f51ae7" containerName="kube-state-metrics" Dec 03 17:39:43 crc kubenswrapper[4787]: E1203 17:39:43.579320 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="561fe2dd-e048-4657-8c0c-2434fe09f35e" containerName="mysqld-exporter" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.579388 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="561fe2dd-e048-4657-8c0c-2434fe09f35e" containerName="mysqld-exporter" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.579758 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5a1dad4-bf28-4c9a-987f-c66658f51ae7" containerName="kube-state-metrics" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.579876 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="561fe2dd-e048-4657-8c0c-2434fe09f35e" containerName="mysqld-exporter" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.581002 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.583502 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.583823 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.593166 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.677497 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.678384 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.678579 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bhjxp\" (UniqueName: \"kubernetes.io/projected/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-kube-api-access-bhjxp\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.678803 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.781100 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.781149 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bhjxp\" (UniqueName: \"kubernetes.io/projected/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-kube-api-access-bhjxp\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.781202 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.781243 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.782993 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5a1dad4-bf28-4c9a-987f-c66658f51ae7" path="/var/lib/kubelet/pods/f5a1dad4-bf28-4c9a-987f-c66658f51ae7/volumes" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.789104 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.793741 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.797630 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.823213 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bhjxp\" (UniqueName: \"kubernetes.io/projected/c3f8fcb4-9cbd-4029-81a8-8cada9c7b305-kube-api-access-bhjxp\") pod \"kube-state-metrics-0\" (UID: \"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305\") " pod="openstack/kube-state-metrics-0" Dec 03 17:39:43 crc kubenswrapper[4787]: I1203 17:39:43.907502 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.231066 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"561fe2dd-e048-4657-8c0c-2434fe09f35e","Type":"ContainerDied","Data":"d138d90d677cba4874a5b2433644142946aa173ee140f87812c4e6174c477742"} Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.231191 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.231568 4787 scope.go:117] "RemoveContainer" containerID="b4de70be6d3d549766d0ae5211c21fd50c58ada0b50b6cdf1f975662db978161" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.274647 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.299565 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.323897 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.325796 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.330069 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"mysqld-exporter-config-data" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.330463 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-mysqld-exporter-svc" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.345260 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.395543 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.395721 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-config-data\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.395787 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-76q4j\" (UniqueName: \"kubernetes.io/projected/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-kube-api-access-76q4j\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.395964 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.475924 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 03 17:39:44 crc kubenswrapper[4787]: W1203 17:39:44.479422 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc3f8fcb4_9cbd_4029_81a8_8cada9c7b305.slice/crio-bae9b19822f465ba481fe9c572fafb797cc4305018de2fe05cd5f66daa9c7575 WatchSource:0}: Error finding container bae9b19822f465ba481fe9c572fafb797cc4305018de2fe05cd5f66daa9c7575: Status 404 returned error can't find the container with id bae9b19822f465ba481fe9c572fafb797cc4305018de2fe05cd5f66daa9c7575 Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.498559 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-config-data\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.498641 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-76q4j\" (UniqueName: \"kubernetes.io/projected/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-kube-api-access-76q4j\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.498744 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.498796 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.512841 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mysqld-exporter-tls-certs\" (UniqueName: \"kubernetes.io/secret/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-mysqld-exporter-tls-certs\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.515461 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-config-data\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.521848 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-76q4j\" (UniqueName: \"kubernetes.io/projected/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-kube-api-access-76q4j\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.535433 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417-combined-ca-bundle\") pod \"mysqld-exporter-0\" (UID: \"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417\") " pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.651540 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/mysqld-exporter-0" Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.736622 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.736902 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="ceilometer-central-agent" containerID="cri-o://4393b805a6c73c8c328ce96a80287c5d9d89704a98d6c82405d251950beafa99" gracePeriod=30 Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.736930 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="sg-core" containerID="cri-o://9d438feea66735f410e7fffbd0a1cfb3b45bedb05623a2b1fed636c3a364b6f1" gracePeriod=30 Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.736977 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="ceilometer-notification-agent" containerID="cri-o://dbe767f04114b5df04bc3a17a2c3c49e5e55b22cb7e8e16c2b6f320d58df3607" gracePeriod=30 Dec 03 17:39:44 crc kubenswrapper[4787]: I1203 17:39:44.736994 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="proxy-httpd" containerID="cri-o://ace71b94e4f14fa0b698938b04697e85cf9cba33cd48f6ffb9680582947c12c3" gracePeriod=30 Dec 03 17:39:45 crc kubenswrapper[4787]: I1203 17:39:45.159141 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/mysqld-exporter-0"] Dec 03 17:39:45 crc kubenswrapper[4787]: W1203 17:39:45.164311 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcd6d5b28_0e8a_4ba4_b6f1_716e6f8e4417.slice/crio-f5f0dfd3450a453d966c10b08b26472dace9cb1fca05647a0554fcd2efdc81f6 WatchSource:0}: Error finding container f5f0dfd3450a453d966c10b08b26472dace9cb1fca05647a0554fcd2efdc81f6: Status 404 returned error can't find the container with id f5f0dfd3450a453d966c10b08b26472dace9cb1fca05647a0554fcd2efdc81f6 Dec 03 17:39:45 crc kubenswrapper[4787]: I1203 17:39:45.249012 4787 generic.go:334] "Generic (PLEG): container finished" podID="f923a880-083f-464a-803a-a4a47662d75f" containerID="ace71b94e4f14fa0b698938b04697e85cf9cba33cd48f6ffb9680582947c12c3" exitCode=0 Dec 03 17:39:45 crc kubenswrapper[4787]: I1203 17:39:45.249067 4787 generic.go:334] "Generic (PLEG): container finished" podID="f923a880-083f-464a-803a-a4a47662d75f" containerID="9d438feea66735f410e7fffbd0a1cfb3b45bedb05623a2b1fed636c3a364b6f1" exitCode=2 Dec 03 17:39:45 crc kubenswrapper[4787]: I1203 17:39:45.249109 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f923a880-083f-464a-803a-a4a47662d75f","Type":"ContainerDied","Data":"ace71b94e4f14fa0b698938b04697e85cf9cba33cd48f6ffb9680582947c12c3"} Dec 03 17:39:45 crc kubenswrapper[4787]: I1203 17:39:45.249141 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f923a880-083f-464a-803a-a4a47662d75f","Type":"ContainerDied","Data":"9d438feea66735f410e7fffbd0a1cfb3b45bedb05623a2b1fed636c3a364b6f1"} Dec 03 17:39:45 crc kubenswrapper[4787]: I1203 17:39:45.251291 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305","Type":"ContainerStarted","Data":"d82753c480b8731fe388410182a8ac8d69d77ecd7ef7f719dbcf7e5af97bf8ee"} Dec 03 17:39:45 crc kubenswrapper[4787]: I1203 17:39:45.251346 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"c3f8fcb4-9cbd-4029-81a8-8cada9c7b305","Type":"ContainerStarted","Data":"bae9b19822f465ba481fe9c572fafb797cc4305018de2fe05cd5f66daa9c7575"} Dec 03 17:39:45 crc kubenswrapper[4787]: I1203 17:39:45.251629 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 03 17:39:45 crc kubenswrapper[4787]: I1203 17:39:45.253365 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417","Type":"ContainerStarted","Data":"f5f0dfd3450a453d966c10b08b26472dace9cb1fca05647a0554fcd2efdc81f6"} Dec 03 17:39:45 crc kubenswrapper[4787]: I1203 17:39:45.276581 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.833959812 podStartE2EDuration="2.276557695s" podCreationTimestamp="2025-12-03 17:39:43 +0000 UTC" firstStartedPulling="2025-12-03 17:39:44.483144391 +0000 UTC m=+1601.300615350" lastFinishedPulling="2025-12-03 17:39:44.925742274 +0000 UTC m=+1601.743213233" observedRunningTime="2025-12-03 17:39:45.269983346 +0000 UTC m=+1602.087454305" watchObservedRunningTime="2025-12-03 17:39:45.276557695 +0000 UTC m=+1602.094028654" Dec 03 17:39:45 crc kubenswrapper[4787]: I1203 17:39:45.780346 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="561fe2dd-e048-4657-8c0c-2434fe09f35e" path="/var/lib/kubelet/pods/561fe2dd-e048-4657-8c0c-2434fe09f35e/volumes" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.277358 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/mysqld-exporter-0" event={"ID":"cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417","Type":"ContainerStarted","Data":"b7a8f5bc41d4728fbb0db072d40a5cbc3ecc789f36a0e958ff16a7d61f9374ae"} Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.284921 4787 generic.go:334] "Generic (PLEG): container finished" podID="f923a880-083f-464a-803a-a4a47662d75f" containerID="dbe767f04114b5df04bc3a17a2c3c49e5e55b22cb7e8e16c2b6f320d58df3607" exitCode=0 Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.284954 4787 generic.go:334] "Generic (PLEG): container finished" podID="f923a880-083f-464a-803a-a4a47662d75f" containerID="4393b805a6c73c8c328ce96a80287c5d9d89704a98d6c82405d251950beafa99" exitCode=0 Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.285013 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f923a880-083f-464a-803a-a4a47662d75f","Type":"ContainerDied","Data":"dbe767f04114b5df04bc3a17a2c3c49e5e55b22cb7e8e16c2b6f320d58df3607"} Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.285119 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f923a880-083f-464a-803a-a4a47662d75f","Type":"ContainerDied","Data":"4393b805a6c73c8c328ce96a80287c5d9d89704a98d6c82405d251950beafa99"} Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.315183 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/mysqld-exporter-0" podStartSLOduration=1.76376062 podStartE2EDuration="2.315158224s" podCreationTimestamp="2025-12-03 17:39:44 +0000 UTC" firstStartedPulling="2025-12-03 17:39:45.167457995 +0000 UTC m=+1601.984928954" lastFinishedPulling="2025-12-03 17:39:45.718855599 +0000 UTC m=+1602.536326558" observedRunningTime="2025-12-03 17:39:46.298848601 +0000 UTC m=+1603.116319560" watchObservedRunningTime="2025-12-03 17:39:46.315158224 +0000 UTC m=+1603.132629183" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.487724 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.547680 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-run-httpd\") pod \"f923a880-083f-464a-803a-a4a47662d75f\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.547779 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-scripts\") pod \"f923a880-083f-464a-803a-a4a47662d75f\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.547808 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-log-httpd\") pod \"f923a880-083f-464a-803a-a4a47662d75f\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.547963 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dp5vg\" (UniqueName: \"kubernetes.io/projected/f923a880-083f-464a-803a-a4a47662d75f-kube-api-access-dp5vg\") pod \"f923a880-083f-464a-803a-a4a47662d75f\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.548032 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-sg-core-conf-yaml\") pod \"f923a880-083f-464a-803a-a4a47662d75f\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.548149 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-config-data\") pod \"f923a880-083f-464a-803a-a4a47662d75f\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.548187 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-combined-ca-bundle\") pod \"f923a880-083f-464a-803a-a4a47662d75f\" (UID: \"f923a880-083f-464a-803a-a4a47662d75f\") " Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.548237 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "f923a880-083f-464a-803a-a4a47662d75f" (UID: "f923a880-083f-464a-803a-a4a47662d75f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.548761 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.549511 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "f923a880-083f-464a-803a-a4a47662d75f" (UID: "f923a880-083f-464a-803a-a4a47662d75f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.554885 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f923a880-083f-464a-803a-a4a47662d75f-kube-api-access-dp5vg" (OuterVolumeSpecName: "kube-api-access-dp5vg") pod "f923a880-083f-464a-803a-a4a47662d75f" (UID: "f923a880-083f-464a-803a-a4a47662d75f"). InnerVolumeSpecName "kube-api-access-dp5vg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.564660 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-scripts" (OuterVolumeSpecName: "scripts") pod "f923a880-083f-464a-803a-a4a47662d75f" (UID: "f923a880-083f-464a-803a-a4a47662d75f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.599849 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "f923a880-083f-464a-803a-a4a47662d75f" (UID: "f923a880-083f-464a-803a-a4a47662d75f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.651077 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dp5vg\" (UniqueName: \"kubernetes.io/projected/f923a880-083f-464a-803a-a4a47662d75f-kube-api-access-dp5vg\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.651106 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.651116 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.651124 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f923a880-083f-464a-803a-a4a47662d75f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.652000 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f923a880-083f-464a-803a-a4a47662d75f" (UID: "f923a880-083f-464a-803a-a4a47662d75f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.687102 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-config-data" (OuterVolumeSpecName: "config-data") pod "f923a880-083f-464a-803a-a4a47662d75f" (UID: "f923a880-083f-464a-803a-a4a47662d75f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.752795 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:46 crc kubenswrapper[4787]: I1203 17:39:46.752987 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f923a880-083f-464a-803a-a4a47662d75f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.296734 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"f923a880-083f-464a-803a-a4a47662d75f","Type":"ContainerDied","Data":"3d02b33644b8efdd4cfe2a93f12a28e05b749e9947c74f01017e5cd54dfe8d93"} Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.296794 4787 scope.go:117] "RemoveContainer" containerID="ace71b94e4f14fa0b698938b04697e85cf9cba33cd48f6ffb9680582947c12c3" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.296803 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.321524 4787 scope.go:117] "RemoveContainer" containerID="9d438feea66735f410e7fffbd0a1cfb3b45bedb05623a2b1fed636c3a364b6f1" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.347868 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.360572 4787 scope.go:117] "RemoveContainer" containerID="dbe767f04114b5df04bc3a17a2c3c49e5e55b22cb7e8e16c2b6f320d58df3607" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.367989 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.388075 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:39:47 crc kubenswrapper[4787]: E1203 17:39:47.388602 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="ceilometer-central-agent" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.388629 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="ceilometer-central-agent" Dec 03 17:39:47 crc kubenswrapper[4787]: E1203 17:39:47.388666 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="proxy-httpd" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.388674 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="proxy-httpd" Dec 03 17:39:47 crc kubenswrapper[4787]: E1203 17:39:47.388690 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="sg-core" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.388699 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="sg-core" Dec 03 17:39:47 crc kubenswrapper[4787]: E1203 17:39:47.388730 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="ceilometer-notification-agent" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.388738 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="ceilometer-notification-agent" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.388998 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="sg-core" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.389048 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="proxy-httpd" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.389064 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="ceilometer-notification-agent" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.389083 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f923a880-083f-464a-803a-a4a47662d75f" containerName="ceilometer-central-agent" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.391357 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.395115 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.395473 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.397601 4787 scope.go:117] "RemoveContainer" containerID="4393b805a6c73c8c328ce96a80287c5d9d89704a98d6c82405d251950beafa99" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.398751 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.400106 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.468308 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.468865 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.468979 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-config-data\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.469009 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-blchw\" (UniqueName: \"kubernetes.io/projected/46f18248-295b-4cae-b29e-51f3cd414799-kube-api-access-blchw\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.469078 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-run-httpd\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.469096 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.469119 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-scripts\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.469149 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-log-httpd\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.570612 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.570711 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.570832 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-config-data\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.570870 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-blchw\" (UniqueName: \"kubernetes.io/projected/46f18248-295b-4cae-b29e-51f3cd414799-kube-api-access-blchw\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.570900 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-run-httpd\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.570927 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.570949 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-scripts\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.570985 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-log-httpd\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.571436 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-log-httpd\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.571437 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-run-httpd\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.575150 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.575342 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.575470 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-config-data\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.575620 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.582861 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-scripts\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.591330 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-blchw\" (UniqueName: \"kubernetes.io/projected/46f18248-295b-4cae-b29e-51f3cd414799-kube-api-access-blchw\") pod \"ceilometer-0\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.711461 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:39:47 crc kubenswrapper[4787]: I1203 17:39:47.785911 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f923a880-083f-464a-803a-a4a47662d75f" path="/var/lib/kubelet/pods/f923a880-083f-464a-803a-a4a47662d75f/volumes" Dec 03 17:39:48 crc kubenswrapper[4787]: W1203 17:39:48.211853 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod46f18248_295b_4cae_b29e_51f3cd414799.slice/crio-e23a3b7a6ee7b31b324dddcb1f6eb752c4a76941c87aff5b0803b35c3c2e3b3c WatchSource:0}: Error finding container e23a3b7a6ee7b31b324dddcb1f6eb752c4a76941c87aff5b0803b35c3c2e3b3c: Status 404 returned error can't find the container with id e23a3b7a6ee7b31b324dddcb1f6eb752c4a76941c87aff5b0803b35c3c2e3b3c Dec 03 17:39:48 crc kubenswrapper[4787]: I1203 17:39:48.220152 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:39:48 crc kubenswrapper[4787]: I1203 17:39:48.307446 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46f18248-295b-4cae-b29e-51f3cd414799","Type":"ContainerStarted","Data":"e23a3b7a6ee7b31b324dddcb1f6eb752c4a76941c87aff5b0803b35c3c2e3b3c"} Dec 03 17:39:49 crc kubenswrapper[4787]: I1203 17:39:49.323582 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46f18248-295b-4cae-b29e-51f3cd414799","Type":"ContainerStarted","Data":"a2d301bdea65ea2d7ba739329f3705cb2e979737f67b885a303b4228a0e47d30"} Dec 03 17:39:50 crc kubenswrapper[4787]: I1203 17:39:50.340281 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46f18248-295b-4cae-b29e-51f3cd414799","Type":"ContainerStarted","Data":"735d1b64d48eb20e4d7012118cb0d88b7c157ba42e9ea77a01e43213f15d5e06"} Dec 03 17:39:50 crc kubenswrapper[4787]: I1203 17:39:50.340817 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46f18248-295b-4cae-b29e-51f3cd414799","Type":"ContainerStarted","Data":"d99e13faf2926475c6d1564399cfd245fc9a3ea4a53d896c9fb137efb2f3aef3"} Dec 03 17:39:52 crc kubenswrapper[4787]: I1203 17:39:52.361090 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46f18248-295b-4cae-b29e-51f3cd414799","Type":"ContainerStarted","Data":"663b5ffc26d19f74661ede6a24c9fefa2a2b98922ed935e6646640326e531b49"} Dec 03 17:39:52 crc kubenswrapper[4787]: I1203 17:39:52.361748 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:39:52 crc kubenswrapper[4787]: I1203 17:39:52.391516 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.5032000720000003 podStartE2EDuration="5.391497664s" podCreationTimestamp="2025-12-03 17:39:47 +0000 UTC" firstStartedPulling="2025-12-03 17:39:48.214825234 +0000 UTC m=+1605.032296203" lastFinishedPulling="2025-12-03 17:39:51.103122836 +0000 UTC m=+1607.920593795" observedRunningTime="2025-12-03 17:39:52.386752315 +0000 UTC m=+1609.204223294" watchObservedRunningTime="2025-12-03 17:39:52.391497664 +0000 UTC m=+1609.208968623" Dec 03 17:39:53 crc kubenswrapper[4787]: I1203 17:39:53.919449 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 03 17:40:17 crc kubenswrapper[4787]: I1203 17:40:17.721004 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 17:40:18 crc kubenswrapper[4787]: I1203 17:40:18.989811 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:40:18 crc kubenswrapper[4787]: I1203 17:40:18.991342 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.047370 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-6pgw8"] Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.059903 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-6pgw8"] Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.171490 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-db-sync-2gfq8"] Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.173098 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2gfq8" Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.186599 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-2gfq8"] Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.267668 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27wt9\" (UniqueName: \"kubernetes.io/projected/5749d651-999d-4d54-8661-1962fc41521c-kube-api-access-27wt9\") pod \"heat-db-sync-2gfq8\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " pod="openstack/heat-db-sync-2gfq8" Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.267721 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-combined-ca-bundle\") pod \"heat-db-sync-2gfq8\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " pod="openstack/heat-db-sync-2gfq8" Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.267916 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-config-data\") pod \"heat-db-sync-2gfq8\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " pod="openstack/heat-db-sync-2gfq8" Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.369465 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-config-data\") pod \"heat-db-sync-2gfq8\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " pod="openstack/heat-db-sync-2gfq8" Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.369613 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27wt9\" (UniqueName: \"kubernetes.io/projected/5749d651-999d-4d54-8661-1962fc41521c-kube-api-access-27wt9\") pod \"heat-db-sync-2gfq8\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " pod="openstack/heat-db-sync-2gfq8" Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.369640 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-combined-ca-bundle\") pod \"heat-db-sync-2gfq8\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " pod="openstack/heat-db-sync-2gfq8" Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.377517 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-config-data\") pod \"heat-db-sync-2gfq8\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " pod="openstack/heat-db-sync-2gfq8" Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.384762 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-combined-ca-bundle\") pod \"heat-db-sync-2gfq8\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " pod="openstack/heat-db-sync-2gfq8" Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.390762 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27wt9\" (UniqueName: \"kubernetes.io/projected/5749d651-999d-4d54-8661-1962fc41521c-kube-api-access-27wt9\") pod \"heat-db-sync-2gfq8\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " pod="openstack/heat-db-sync-2gfq8" Dec 03 17:40:30 crc kubenswrapper[4787]: I1203 17:40:30.507982 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2gfq8" Dec 03 17:40:31 crc kubenswrapper[4787]: I1203 17:40:31.010090 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-db-sync-2gfq8"] Dec 03 17:40:31 crc kubenswrapper[4787]: W1203 17:40:31.027526 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5749d651_999d_4d54_8661_1962fc41521c.slice/crio-32882de188c35d334659b6e473f4f394a1c53cf018ef993169059cd100552012 WatchSource:0}: Error finding container 32882de188c35d334659b6e473f4f394a1c53cf018ef993169059cd100552012: Status 404 returned error can't find the container with id 32882de188c35d334659b6e473f4f394a1c53cf018ef993169059cd100552012 Dec 03 17:40:31 crc kubenswrapper[4787]: I1203 17:40:31.031496 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:40:31 crc kubenswrapper[4787]: I1203 17:40:31.780937 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e5395cf-a94f-4125-8d92-54b7cb9da3be" path="/var/lib/kubelet/pods/5e5395cf-a94f-4125-8d92-54b7cb9da3be/volumes" Dec 03 17:40:31 crc kubenswrapper[4787]: I1203 17:40:31.824970 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2gfq8" event={"ID":"5749d651-999d-4d54-8661-1962fc41521c","Type":"ContainerStarted","Data":"32882de188c35d334659b6e473f4f394a1c53cf018ef993169059cd100552012"} Dec 03 17:40:32 crc kubenswrapper[4787]: I1203 17:40:32.497571 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:40:32 crc kubenswrapper[4787]: I1203 17:40:32.667591 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:40:32 crc kubenswrapper[4787]: I1203 17:40:32.668449 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="ceilometer-central-agent" containerID="cri-o://a2d301bdea65ea2d7ba739329f3705cb2e979737f67b885a303b4228a0e47d30" gracePeriod=30 Dec 03 17:40:32 crc kubenswrapper[4787]: I1203 17:40:32.668925 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="proxy-httpd" containerID="cri-o://663b5ffc26d19f74661ede6a24c9fefa2a2b98922ed935e6646640326e531b49" gracePeriod=30 Dec 03 17:40:32 crc kubenswrapper[4787]: I1203 17:40:32.668978 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="sg-core" containerID="cri-o://735d1b64d48eb20e4d7012118cb0d88b7c157ba42e9ea77a01e43213f15d5e06" gracePeriod=30 Dec 03 17:40:32 crc kubenswrapper[4787]: I1203 17:40:32.669010 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="ceilometer-notification-agent" containerID="cri-o://d99e13faf2926475c6d1564399cfd245fc9a3ea4a53d896c9fb137efb2f3aef3" gracePeriod=30 Dec 03 17:40:33 crc kubenswrapper[4787]: I1203 17:40:33.502310 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:40:33 crc kubenswrapper[4787]: I1203 17:40:33.865897 4787 generic.go:334] "Generic (PLEG): container finished" podID="46f18248-295b-4cae-b29e-51f3cd414799" containerID="663b5ffc26d19f74661ede6a24c9fefa2a2b98922ed935e6646640326e531b49" exitCode=0 Dec 03 17:40:33 crc kubenswrapper[4787]: I1203 17:40:33.865957 4787 generic.go:334] "Generic (PLEG): container finished" podID="46f18248-295b-4cae-b29e-51f3cd414799" containerID="735d1b64d48eb20e4d7012118cb0d88b7c157ba42e9ea77a01e43213f15d5e06" exitCode=2 Dec 03 17:40:33 crc kubenswrapper[4787]: I1203 17:40:33.865973 4787 generic.go:334] "Generic (PLEG): container finished" podID="46f18248-295b-4cae-b29e-51f3cd414799" containerID="a2d301bdea65ea2d7ba739329f3705cb2e979737f67b885a303b4228a0e47d30" exitCode=0 Dec 03 17:40:33 crc kubenswrapper[4787]: I1203 17:40:33.866044 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46f18248-295b-4cae-b29e-51f3cd414799","Type":"ContainerDied","Data":"663b5ffc26d19f74661ede6a24c9fefa2a2b98922ed935e6646640326e531b49"} Dec 03 17:40:33 crc kubenswrapper[4787]: I1203 17:40:33.866088 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46f18248-295b-4cae-b29e-51f3cd414799","Type":"ContainerDied","Data":"735d1b64d48eb20e4d7012118cb0d88b7c157ba42e9ea77a01e43213f15d5e06"} Dec 03 17:40:33 crc kubenswrapper[4787]: I1203 17:40:33.866128 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46f18248-295b-4cae-b29e-51f3cd414799","Type":"ContainerDied","Data":"a2d301bdea65ea2d7ba739329f3705cb2e979737f67b885a303b4228a0e47d30"} Dec 03 17:40:37 crc kubenswrapper[4787]: I1203 17:40:37.939639 4787 generic.go:334] "Generic (PLEG): container finished" podID="46f18248-295b-4cae-b29e-51f3cd414799" containerID="d99e13faf2926475c6d1564399cfd245fc9a3ea4a53d896c9fb137efb2f3aef3" exitCode=0 Dec 03 17:40:37 crc kubenswrapper[4787]: I1203 17:40:37.939708 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46f18248-295b-4cae-b29e-51f3cd414799","Type":"ContainerDied","Data":"d99e13faf2926475c6d1564399cfd245fc9a3ea4a53d896c9fb137efb2f3aef3"} Dec 03 17:40:38 crc kubenswrapper[4787]: I1203 17:40:38.121685 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="d73396bd-2e1f-4c20-8327-610c5d783373" containerName="rabbitmq" containerID="cri-o://cb713dd212ea4276428258b4b48c6b045aa425d46ef7ca7d7e569df15f21de58" gracePeriod=604795 Dec 03 17:40:38 crc kubenswrapper[4787]: I1203 17:40:38.510839 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="3180168d-9dad-441d-86bd-9d1801ef629f" containerName="rabbitmq" containerID="cri-o://b4744a19f52cb6b0397c6428dd54a30ebe8f002f3d0e4df0a4d07e3e17f6fea0" gracePeriod=604795 Dec 03 17:40:44 crc kubenswrapper[4787]: I1203 17:40:44.044342 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="3180168d-9dad-441d-86bd-9d1801ef629f" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.119:5671: connect: connection refused" Dec 03 17:40:44 crc kubenswrapper[4787]: I1203 17:40:44.355148 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="d73396bd-2e1f-4c20-8327-610c5d783373" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.120:5671: connect: connection refused" Dec 03 17:40:45 crc kubenswrapper[4787]: I1203 17:40:45.022216 4787 generic.go:334] "Generic (PLEG): container finished" podID="d73396bd-2e1f-4c20-8327-610c5d783373" containerID="cb713dd212ea4276428258b4b48c6b045aa425d46ef7ca7d7e569df15f21de58" exitCode=0 Dec 03 17:40:45 crc kubenswrapper[4787]: I1203 17:40:45.022298 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d73396bd-2e1f-4c20-8327-610c5d783373","Type":"ContainerDied","Data":"cb713dd212ea4276428258b4b48c6b045aa425d46ef7ca7d7e569df15f21de58"} Dec 03 17:40:45 crc kubenswrapper[4787]: I1203 17:40:45.025665 4787 generic.go:334] "Generic (PLEG): container finished" podID="3180168d-9dad-441d-86bd-9d1801ef629f" containerID="b4744a19f52cb6b0397c6428dd54a30ebe8f002f3d0e4df0a4d07e3e17f6fea0" exitCode=0 Dec 03 17:40:45 crc kubenswrapper[4787]: I1203 17:40:45.025755 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3180168d-9dad-441d-86bd-9d1801ef629f","Type":"ContainerDied","Data":"b4744a19f52cb6b0397c6428dd54a30ebe8f002f3d0e4df0a4d07e3e17f6fea0"} Dec 03 17:40:48 crc kubenswrapper[4787]: I1203 17:40:48.989450 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:40:48 crc kubenswrapper[4787]: I1203 17:40:48.990224 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:40:50 crc kubenswrapper[4787]: E1203 17:40:50.428460 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 03 17:40:50 crc kubenswrapper[4787]: E1203 17:40:50.429435 4787 kuberuntime_image.go:55] "Failed to pull image" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested" Dec 03 17:40:50 crc kubenswrapper[4787]: E1203 17:40:50.429618 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:heat-db-sync,Image:quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested,Command:[/bin/bash],Args:[-c /usr/bin/heat-manage --config-dir /etc/heat/heat.conf.d db_sync],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:true,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/heat/heat.conf.d/00-default.conf,SubPath:00-default.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/heat/heat.conf.d/01-custom.conf,SubPath:01-custom.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-27wt9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42418,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:*42418,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-db-sync-2gfq8_openstack(5749d651-999d-4d54-8661-1962fc41521c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 17:40:50 crc kubenswrapper[4787]: E1203 17:40:50.430845 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/heat-db-sync-2gfq8" podUID="5749d651-999d-4d54-8661-1962fc41521c" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.438858 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-nstbb"] Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.441362 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.445369 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.453102 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-nstbb"] Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.508176 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.516230 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.526863 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.573250 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-config\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.573319 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.573335 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.573383 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.573451 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.573474 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.573509 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-frlth\" (UniqueName: \"kubernetes.io/projected/b1660890-8f51-407f-ba9d-c55853e01702-kube-api-access-frlth\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.674466 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-ceilometer-tls-certs\") pod \"46f18248-295b-4cae-b29e-51f3cd414799\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.675409 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-config-data\") pod \"3180168d-9dad-441d-86bd-9d1801ef629f\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.675534 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-blchw\" (UniqueName: \"kubernetes.io/projected/46f18248-295b-4cae-b29e-51f3cd414799-kube-api-access-blchw\") pod \"46f18248-295b-4cae-b29e-51f3cd414799\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.675612 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5lg9b\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-kube-api-access-5lg9b\") pod \"d73396bd-2e1f-4c20-8327-610c5d783373\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.675734 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-plugins-conf\") pod \"3180168d-9dad-441d-86bd-9d1801ef629f\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.675826 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3180168d-9dad-441d-86bd-9d1801ef629f-erlang-cookie-secret\") pod \"3180168d-9dad-441d-86bd-9d1801ef629f\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.675913 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-plugins\") pod \"3180168d-9dad-441d-86bd-9d1801ef629f\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.676009 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3180168d-9dad-441d-86bd-9d1801ef629f-pod-info\") pod \"3180168d-9dad-441d-86bd-9d1801ef629f\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.676180 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d73396bd-2e1f-4c20-8327-610c5d783373-pod-info\") pod \"d73396bd-2e1f-4c20-8327-610c5d783373\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.676265 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-confd\") pod \"3180168d-9dad-441d-86bd-9d1801ef629f\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.676349 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-config-data\") pod \"46f18248-295b-4cae-b29e-51f3cd414799\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.676435 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-combined-ca-bundle\") pod \"46f18248-295b-4cae-b29e-51f3cd414799\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.676547 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s5b5k\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-kube-api-access-s5b5k\") pod \"3180168d-9dad-441d-86bd-9d1801ef629f\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.676623 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-run-httpd\") pod \"46f18248-295b-4cae-b29e-51f3cd414799\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.676710 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-config-data\") pod \"d73396bd-2e1f-4c20-8327-610c5d783373\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.676824 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-erlang-cookie\") pod \"d73396bd-2e1f-4c20-8327-610c5d783373\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.676915 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-plugins\") pod \"d73396bd-2e1f-4c20-8327-610c5d783373\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.677175 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-server-conf\") pod \"3180168d-9dad-441d-86bd-9d1801ef629f\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.677284 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-tls\") pod \"3180168d-9dad-441d-86bd-9d1801ef629f\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.677367 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-sg-core-conf-yaml\") pod \"46f18248-295b-4cae-b29e-51f3cd414799\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.677453 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-plugins-conf\") pod \"d73396bd-2e1f-4c20-8327-610c5d783373\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.677598 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-tls\") pod \"d73396bd-2e1f-4c20-8327-610c5d783373\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.677712 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-erlang-cookie\") pod \"3180168d-9dad-441d-86bd-9d1801ef629f\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.677816 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-scripts\") pod \"46f18248-295b-4cae-b29e-51f3cd414799\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.677919 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-confd\") pod \"d73396bd-2e1f-4c20-8327-610c5d783373\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.678068 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"3180168d-9dad-441d-86bd-9d1801ef629f\" (UID: \"3180168d-9dad-441d-86bd-9d1801ef629f\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.678168 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"d73396bd-2e1f-4c20-8327-610c5d783373\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.678304 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-server-conf\") pod \"d73396bd-2e1f-4c20-8327-610c5d783373\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.678430 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-log-httpd\") pod \"46f18248-295b-4cae-b29e-51f3cd414799\" (UID: \"46f18248-295b-4cae-b29e-51f3cd414799\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.678661 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d73396bd-2e1f-4c20-8327-610c5d783373-erlang-cookie-secret\") pod \"d73396bd-2e1f-4c20-8327-610c5d783373\" (UID: \"d73396bd-2e1f-4c20-8327-610c5d783373\") " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.679005 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.679630 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.679779 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.679925 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-frlth\" (UniqueName: \"kubernetes.io/projected/b1660890-8f51-407f-ba9d-c55853e01702-kube-api-access-frlth\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.680148 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-config\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.680325 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.680429 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.683139 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-openstack-edpm-ipam\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.676547 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3180168d-9dad-441d-86bd-9d1801ef629f" (UID: "3180168d-9dad-441d-86bd-9d1801ef629f"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.679969 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "d73396bd-2e1f-4c20-8327-610c5d783373" (UID: "d73396bd-2e1f-4c20-8327-610c5d783373"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.680261 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "46f18248-295b-4cae-b29e-51f3cd414799" (UID: "46f18248-295b-4cae-b29e-51f3cd414799"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.680327 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46f18248-295b-4cae-b29e-51f3cd414799-kube-api-access-blchw" (OuterVolumeSpecName: "kube-api-access-blchw") pod "46f18248-295b-4cae-b29e-51f3cd414799" (UID: "46f18248-295b-4cae-b29e-51f3cd414799"). InnerVolumeSpecName "kube-api-access-blchw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.680400 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-kube-api-access-5lg9b" (OuterVolumeSpecName: "kube-api-access-5lg9b") pod "d73396bd-2e1f-4c20-8327-610c5d783373" (UID: "d73396bd-2e1f-4c20-8327-610c5d783373"). InnerVolumeSpecName "kube-api-access-5lg9b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.682437 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3180168d-9dad-441d-86bd-9d1801ef629f-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3180168d-9dad-441d-86bd-9d1801ef629f" (UID: "3180168d-9dad-441d-86bd-9d1801ef629f"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.682681 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/d73396bd-2e1f-4c20-8327-610c5d783373-pod-info" (OuterVolumeSpecName: "pod-info") pod "d73396bd-2e1f-4c20-8327-610c5d783373" (UID: "d73396bd-2e1f-4c20-8327-610c5d783373"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.682724 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-kube-api-access-s5b5k" (OuterVolumeSpecName: "kube-api-access-s5b5k") pod "3180168d-9dad-441d-86bd-9d1801ef629f" (UID: "3180168d-9dad-441d-86bd-9d1801ef629f"). InnerVolumeSpecName "kube-api-access-s5b5k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.683143 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "46f18248-295b-4cae-b29e-51f3cd414799" (UID: "46f18248-295b-4cae-b29e-51f3cd414799"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.687574 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3180168d-9dad-441d-86bd-9d1801ef629f" (UID: "3180168d-9dad-441d-86bd-9d1801ef629f"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.688306 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "3180168d-9dad-441d-86bd-9d1801ef629f" (UID: "3180168d-9dad-441d-86bd-9d1801ef629f"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.688487 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-nb\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.689157 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-svc\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.689768 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-config\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.683599 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-sb\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.690432 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "d73396bd-2e1f-4c20-8327-610c5d783373" (UID: "d73396bd-2e1f-4c20-8327-610c5d783373"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.690607 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d73396bd-2e1f-4c20-8327-610c5d783373-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "d73396bd-2e1f-4c20-8327-610c5d783373" (UID: "d73396bd-2e1f-4c20-8327-610c5d783373"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.691517 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-swift-storage-0\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.692070 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "d73396bd-2e1f-4c20-8327-610c5d783373" (UID: "d73396bd-2e1f-4c20-8327-610c5d783373"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.696368 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3180168d-9dad-441d-86bd-9d1801ef629f" (UID: "3180168d-9dad-441d-86bd-9d1801ef629f"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.696478 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "d73396bd-2e1f-4c20-8327-610c5d783373" (UID: "d73396bd-2e1f-4c20-8327-610c5d783373"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.697707 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3180168d-9dad-441d-86bd-9d1801ef629f-pod-info" (OuterVolumeSpecName: "pod-info") pod "3180168d-9dad-441d-86bd-9d1801ef629f" (UID: "3180168d-9dad-441d-86bd-9d1801ef629f"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.729939 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "persistence") pod "3180168d-9dad-441d-86bd-9d1801ef629f" (UID: "3180168d-9dad-441d-86bd-9d1801ef629f"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.733957 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-scripts" (OuterVolumeSpecName: "scripts") pod "46f18248-295b-4cae-b29e-51f3cd414799" (UID: "46f18248-295b-4cae-b29e-51f3cd414799"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.734331 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "d73396bd-2e1f-4c20-8327-610c5d783373" (UID: "d73396bd-2e1f-4c20-8327-610c5d783373"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.746912 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-frlth\" (UniqueName: \"kubernetes.io/projected/b1660890-8f51-407f-ba9d-c55853e01702-kube-api-access-frlth\") pod \"dnsmasq-dns-7d84b4d45c-nstbb\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793381 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793420 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793432 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793444 4787 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793455 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793466 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793476 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793508 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793524 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793535 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793546 4787 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d73396bd-2e1f-4c20-8327-610c5d783373-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793557 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-blchw\" (UniqueName: \"kubernetes.io/projected/46f18248-295b-4cae-b29e-51f3cd414799-kube-api-access-blchw\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793576 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5lg9b\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-kube-api-access-5lg9b\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793587 4787 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793600 4787 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3180168d-9dad-441d-86bd-9d1801ef629f-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793611 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793621 4787 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3180168d-9dad-441d-86bd-9d1801ef629f-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793631 4787 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d73396bd-2e1f-4c20-8327-610c5d783373-pod-info\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793642 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s5b5k\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-kube-api-access-s5b5k\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.793652 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/46f18248-295b-4cae-b29e-51f3cd414799-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.825230 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "46f18248-295b-4cae-b29e-51f3cd414799" (UID: "46f18248-295b-4cae-b29e-51f3cd414799"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.832385 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.838012 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-config-data" (OuterVolumeSpecName: "config-data") pod "3180168d-9dad-441d-86bd-9d1801ef629f" (UID: "3180168d-9dad-441d-86bd-9d1801ef629f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.850345 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.863555 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-config-data" (OuterVolumeSpecName: "config-data") pod "d73396bd-2e1f-4c20-8327-610c5d783373" (UID: "d73396bd-2e1f-4c20-8327-610c5d783373"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.874579 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "46f18248-295b-4cae-b29e-51f3cd414799" (UID: "46f18248-295b-4cae-b29e-51f3cd414799"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.884451 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.902965 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.903000 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.903027 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.903039 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.903050 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.903060 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.922172 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-server-conf" (OuterVolumeSpecName: "server-conf") pod "3180168d-9dad-441d-86bd-9d1801ef629f" (UID: "3180168d-9dad-441d-86bd-9d1801ef629f"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.928735 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-server-conf" (OuterVolumeSpecName: "server-conf") pod "d73396bd-2e1f-4c20-8327-610c5d783373" (UID: "d73396bd-2e1f-4c20-8327-610c5d783373"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:40:50 crc kubenswrapper[4787]: I1203 17:40:50.997203 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "46f18248-295b-4cae-b29e-51f3cd414799" (UID: "46f18248-295b-4cae-b29e-51f3cd414799"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.016710 4787 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3180168d-9dad-441d-86bd-9d1801ef629f-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.016746 4787 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d73396bd-2e1f-4c20-8327-610c5d783373-server-conf\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.016758 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.037483 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3180168d-9dad-441d-86bd-9d1801ef629f" (UID: "3180168d-9dad-441d-86bd-9d1801ef629f"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.041176 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-config-data" (OuterVolumeSpecName: "config-data") pod "46f18248-295b-4cae-b29e-51f3cd414799" (UID: "46f18248-295b-4cae-b29e-51f3cd414799"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.041244 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "d73396bd-2e1f-4c20-8327-610c5d783373" (UID: "d73396bd-2e1f-4c20-8327-610c5d783373"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.097719 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"46f18248-295b-4cae-b29e-51f3cd414799","Type":"ContainerDied","Data":"e23a3b7a6ee7b31b324dddcb1f6eb752c4a76941c87aff5b0803b35c3c2e3b3c"} Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.097786 4787 scope.go:117] "RemoveContainer" containerID="663b5ffc26d19f74661ede6a24c9fefa2a2b98922ed935e6646640326e531b49" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.097969 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.111390 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.121072 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"3180168d-9dad-441d-86bd-9d1801ef629f","Type":"ContainerDied","Data":"e562f65d2903616d598009a52a3d99823bf9bd894c41b5b25dfb0a9ca535e634"} Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.127294 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d73396bd-2e1f-4c20-8327-610c5d783373-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.127454 4787 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3180168d-9dad-441d-86bd-9d1801ef629f-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.127467 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/46f18248-295b-4cae-b29e-51f3cd414799-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.128344 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.129241 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d73396bd-2e1f-4c20-8327-610c5d783373","Type":"ContainerDied","Data":"0cba0f8daa821c71af0478f2451d263670606b44b0ba3acb56917d287324d6aa"} Dec 03 17:40:51 crc kubenswrapper[4787]: E1203 17:40:51.132315 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"heat-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.rdoproject.org/podified-master-centos10/openstack-heat-engine:current-tested\\\"\"" pod="openstack/heat-db-sync-2gfq8" podUID="5749d651-999d-4d54-8661-1962fc41521c" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.187309 4787 scope.go:117] "RemoveContainer" containerID="735d1b64d48eb20e4d7012118cb0d88b7c157ba42e9ea77a01e43213f15d5e06" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.287220 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.301372 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.315468 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: E1203 17:40:51.316466 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d73396bd-2e1f-4c20-8327-610c5d783373" containerName="rabbitmq" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316494 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d73396bd-2e1f-4c20-8327-610c5d783373" containerName="rabbitmq" Dec 03 17:40:51 crc kubenswrapper[4787]: E1203 17:40:51.316519 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3180168d-9dad-441d-86bd-9d1801ef629f" containerName="setup-container" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316527 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3180168d-9dad-441d-86bd-9d1801ef629f" containerName="setup-container" Dec 03 17:40:51 crc kubenswrapper[4787]: E1203 17:40:51.316541 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="proxy-httpd" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316549 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="proxy-httpd" Dec 03 17:40:51 crc kubenswrapper[4787]: E1203 17:40:51.316569 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="ceilometer-central-agent" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316578 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="ceilometer-central-agent" Dec 03 17:40:51 crc kubenswrapper[4787]: E1203 17:40:51.316596 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d73396bd-2e1f-4c20-8327-610c5d783373" containerName="setup-container" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316604 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d73396bd-2e1f-4c20-8327-610c5d783373" containerName="setup-container" Dec 03 17:40:51 crc kubenswrapper[4787]: E1203 17:40:51.316622 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3180168d-9dad-441d-86bd-9d1801ef629f" containerName="rabbitmq" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316630 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3180168d-9dad-441d-86bd-9d1801ef629f" containerName="rabbitmq" Dec 03 17:40:51 crc kubenswrapper[4787]: E1203 17:40:51.316643 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="sg-core" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316650 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="sg-core" Dec 03 17:40:51 crc kubenswrapper[4787]: E1203 17:40:51.316659 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="ceilometer-notification-agent" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316666 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="ceilometer-notification-agent" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316874 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="proxy-httpd" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316892 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="sg-core" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316901 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="ceilometer-central-agent" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316911 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d73396bd-2e1f-4c20-8327-610c5d783373" containerName="rabbitmq" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316918 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="3180168d-9dad-441d-86bd-9d1801ef629f" containerName="rabbitmq" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.316936 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="ceilometer-notification-agent" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.320327 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.324608 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.325009 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.325182 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.325994 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.327354 4787 scope.go:117] "RemoveContainer" containerID="d99e13faf2926475c6d1564399cfd245fc9a3ea4a53d896c9fb137efb2f3aef3" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.358111 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.371632 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.376435 4787 scope.go:117] "RemoveContainer" containerID="a2d301bdea65ea2d7ba739329f3705cb2e979737f67b885a303b4228a0e47d30" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.382232 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.392552 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.397490 4787 scope.go:117] "RemoveContainer" containerID="b4744a19f52cb6b0397c6428dd54a30ebe8f002f3d0e4df0a4d07e3e17f6fea0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.402944 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.405067 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.410628 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.411728 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.411809 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.413571 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.413700 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.413804 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.413964 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.414208 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-6rzx9" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.433714 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.436938 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.437029 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-run-httpd\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.437053 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-config-data\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.437085 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.437154 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6w84c\" (UniqueName: \"kubernetes.io/projected/1c5f585a-69da-4588-ab44-bc08513f4c9f-kube-api-access-6w84c\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.437204 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-log-httpd\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.437248 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-scripts\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.437265 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.441151 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.445924 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.446376 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.446405 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.446652 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.446927 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.447770 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.448478 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.449260 4787 scope.go:117] "RemoveContainer" containerID="2cad526e0647b21c366d8989df027fd6a0351ec3a648a935c487d2d379ae49af" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.451546 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-mpwfm" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.471677 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-nstbb"] Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.478767 4787 scope.go:117] "RemoveContainer" containerID="cb713dd212ea4276428258b4b48c6b045aa425d46ef7ca7d7e569df15f21de58" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.506142 4787 scope.go:117] "RemoveContainer" containerID="e140a21aafabe034cbd6828830bc11cbed6f27b23fdfcc9449ef640e44a0560b" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539059 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwb4s\" (UniqueName: \"kubernetes.io/projected/9b8eef40-19d4-4640-9517-f603e62e646f-kube-api-access-cwb4s\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539126 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539196 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9b8eef40-19d4-4640-9517-f603e62e646f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539220 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d774ea58-6127-497a-8948-104e60bfe29b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539283 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539312 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d774ea58-6127-497a-8948-104e60bfe29b-config-data\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539348 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539379 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d774ea58-6127-497a-8948-104e60bfe29b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539403 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539433 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539469 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539524 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-run-httpd\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539549 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-config-data\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539581 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539604 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9b8eef40-19d4-4640-9517-f603e62e646f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539644 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539668 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d774ea58-6127-497a-8948-104e60bfe29b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539701 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9b8eef40-19d4-4640-9517-f603e62e646f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539739 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6w84c\" (UniqueName: \"kubernetes.io/projected/1c5f585a-69da-4588-ab44-bc08513f4c9f-kube-api-access-6w84c\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539765 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539806 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b8eef40-19d4-4640-9517-f603e62e646f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539829 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d774ea58-6127-497a-8948-104e60bfe29b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539859 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539885 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-log-httpd\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539913 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539932 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539966 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9b8eef40-19d4-4640-9517-f603e62e646f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.539995 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-scripts\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.540032 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.540066 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7vn7t\" (UniqueName: \"kubernetes.io/projected/d774ea58-6127-497a-8948-104e60bfe29b-kube-api-access-7vn7t\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.540812 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-run-httpd\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.541000 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-log-httpd\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.543925 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-scripts\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.545468 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.545568 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.545735 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.546735 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-config-data\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.562163 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6w84c\" (UniqueName: \"kubernetes.io/projected/1c5f585a-69da-4588-ab44-bc08513f4c9f-kube-api-access-6w84c\") pod \"ceilometer-0\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642302 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9b8eef40-19d4-4640-9517-f603e62e646f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642369 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642420 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b8eef40-19d4-4640-9517-f603e62e646f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642445 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d774ea58-6127-497a-8948-104e60bfe29b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642476 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642535 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642561 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642598 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9b8eef40-19d4-4640-9517-f603e62e646f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642638 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7vn7t\" (UniqueName: \"kubernetes.io/projected/d774ea58-6127-497a-8948-104e60bfe29b-kube-api-access-7vn7t\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642675 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwb4s\" (UniqueName: \"kubernetes.io/projected/9b8eef40-19d4-4640-9517-f603e62e646f-kube-api-access-cwb4s\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642707 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642731 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9b8eef40-19d4-4640-9517-f603e62e646f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642755 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d774ea58-6127-497a-8948-104e60bfe29b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642785 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642813 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d774ea58-6127-497a-8948-104e60bfe29b-config-data\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642851 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642882 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d774ea58-6127-497a-8948-104e60bfe29b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642903 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642947 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.642974 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.643009 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9b8eef40-19d4-4640-9517-f603e62e646f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.643041 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.643077 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.643103 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d774ea58-6127-497a-8948-104e60bfe29b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.643415 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.643512 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9b8eef40-19d4-4640-9517-f603e62e646f-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.643533 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d774ea58-6127-497a-8948-104e60bfe29b-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.643688 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/9b8eef40-19d4-4640-9517-f603e62e646f-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.644007 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.644037 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/9b8eef40-19d4-4640-9517-f603e62e646f-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.644212 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.645115 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/d774ea58-6127-497a-8948-104e60bfe29b-server-conf\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.645401 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.647034 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/d774ea58-6127-497a-8948-104e60bfe29b-config-data\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.647292 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.647804 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/9b8eef40-19d4-4640-9517-f603e62e646f-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.648505 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d774ea58-6127-497a-8948-104e60bfe29b-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.650061 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.650202 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/d774ea58-6127-497a-8948-104e60bfe29b-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.653619 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d774ea58-6127-497a-8948-104e60bfe29b-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.653840 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/9b8eef40-19d4-4640-9517-f603e62e646f-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.655288 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/9b8eef40-19d4-4640-9517-f603e62e646f-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.656335 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.669740 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwb4s\" (UniqueName: \"kubernetes.io/projected/9b8eef40-19d4-4640-9517-f603e62e646f-kube-api-access-cwb4s\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.670078 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7vn7t\" (UniqueName: \"kubernetes.io/projected/d774ea58-6127-497a-8948-104e60bfe29b-kube-api-access-7vn7t\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.709114 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"9b8eef40-19d4-4640-9517-f603e62e646f\") " pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.721251 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-server-0\" (UID: \"d774ea58-6127-497a-8948-104e60bfe29b\") " pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.736794 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.763341 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.798085 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3180168d-9dad-441d-86bd-9d1801ef629f" path="/var/lib/kubelet/pods/3180168d-9dad-441d-86bd-9d1801ef629f/volumes" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.799619 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46f18248-295b-4cae-b29e-51f3cd414799" path="/var/lib/kubelet/pods/46f18248-295b-4cae-b29e-51f3cd414799/volumes" Dec 03 17:40:51 crc kubenswrapper[4787]: I1203 17:40:51.801714 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d73396bd-2e1f-4c20-8327-610c5d783373" path="/var/lib/kubelet/pods/d73396bd-2e1f-4c20-8327-610c5d783373/volumes" Dec 03 17:40:52 crc kubenswrapper[4787]: I1203 17:40:52.205345 4787 generic.go:334] "Generic (PLEG): container finished" podID="b1660890-8f51-407f-ba9d-c55853e01702" containerID="4cd58953ea8fe9ca6599b8d1eebbcd8d5d6b6c22b3b4f490e842326d9184327f" exitCode=0 Dec 03 17:40:52 crc kubenswrapper[4787]: I1203 17:40:52.205672 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" event={"ID":"b1660890-8f51-407f-ba9d-c55853e01702","Type":"ContainerDied","Data":"4cd58953ea8fe9ca6599b8d1eebbcd8d5d6b6c22b3b4f490e842326d9184327f"} Dec 03 17:40:52 crc kubenswrapper[4787]: I1203 17:40:52.205709 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" event={"ID":"b1660890-8f51-407f-ba9d-c55853e01702","Type":"ContainerStarted","Data":"178844243ac3037e1fa8448e59f097d84d4acee48fd89bc173968af840dbbd1b"} Dec 03 17:40:52 crc kubenswrapper[4787]: I1203 17:40:52.307087 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 17:40:52 crc kubenswrapper[4787]: W1203 17:40:52.532635 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd774ea58_6127_497a_8948_104e60bfe29b.slice/crio-8eabca8804149679b210f06293d2e67edfd4d33f419152d7d289b23ec41efc6d WatchSource:0}: Error finding container 8eabca8804149679b210f06293d2e67edfd4d33f419152d7d289b23ec41efc6d: Status 404 returned error can't find the container with id 8eabca8804149679b210f06293d2e67edfd4d33f419152d7d289b23ec41efc6d Dec 03 17:40:52 crc kubenswrapper[4787]: I1203 17:40:52.536232 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 03 17:40:52 crc kubenswrapper[4787]: W1203 17:40:52.538515 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b8eef40_19d4_4640_9517_f603e62e646f.slice/crio-3a394249fff7f47bc6cbe4bbdd7894d758f3c225ed6517eb7f602a95dfdd43e0 WatchSource:0}: Error finding container 3a394249fff7f47bc6cbe4bbdd7894d758f3c225ed6517eb7f602a95dfdd43e0: Status 404 returned error can't find the container with id 3a394249fff7f47bc6cbe4bbdd7894d758f3c225ed6517eb7f602a95dfdd43e0 Dec 03 17:40:52 crc kubenswrapper[4787]: I1203 17:40:52.551257 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 03 17:40:53 crc kubenswrapper[4787]: I1203 17:40:53.237544 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9b8eef40-19d4-4640-9517-f603e62e646f","Type":"ContainerStarted","Data":"3a394249fff7f47bc6cbe4bbdd7894d758f3c225ed6517eb7f602a95dfdd43e0"} Dec 03 17:40:53 crc kubenswrapper[4787]: I1203 17:40:53.243410 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d774ea58-6127-497a-8948-104e60bfe29b","Type":"ContainerStarted","Data":"8eabca8804149679b210f06293d2e67edfd4d33f419152d7d289b23ec41efc6d"} Dec 03 17:40:53 crc kubenswrapper[4787]: I1203 17:40:53.252066 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c5f585a-69da-4588-ab44-bc08513f4c9f","Type":"ContainerStarted","Data":"0827e4d2014acf3c689ae6906dfcb5a72c7c820a60270194027f058e645ee732"} Dec 03 17:40:53 crc kubenswrapper[4787]: I1203 17:40:53.256413 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" event={"ID":"b1660890-8f51-407f-ba9d-c55853e01702","Type":"ContainerStarted","Data":"3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643"} Dec 03 17:40:53 crc kubenswrapper[4787]: I1203 17:40:53.256630 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:40:53 crc kubenswrapper[4787]: I1203 17:40:53.278778 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" podStartSLOduration=3.278756612 podStartE2EDuration="3.278756612s" podCreationTimestamp="2025-12-03 17:40:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:40:53.273410007 +0000 UTC m=+1670.090880986" watchObservedRunningTime="2025-12-03 17:40:53.278756612 +0000 UTC m=+1670.096227571" Dec 03 17:40:55 crc kubenswrapper[4787]: I1203 17:40:55.286419 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d774ea58-6127-497a-8948-104e60bfe29b","Type":"ContainerStarted","Data":"a6d6eaceb90dc23604c321c6f66e5ec95d1e94f8a7599bfec0610dba36784ae9"} Dec 03 17:40:55 crc kubenswrapper[4787]: I1203 17:40:55.288917 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9b8eef40-19d4-4640-9517-f603e62e646f","Type":"ContainerStarted","Data":"52e28b1084238d8e9e3cb69d554cf7061d54c2b4d3d74e258d7b241ad1274498"} Dec 03 17:40:57 crc kubenswrapper[4787]: I1203 17:40:57.313250 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c5f585a-69da-4588-ab44-bc08513f4c9f","Type":"ContainerStarted","Data":"4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c"} Dec 03 17:40:57 crc kubenswrapper[4787]: I1203 17:40:57.313915 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c5f585a-69da-4588-ab44-bc08513f4c9f","Type":"ContainerStarted","Data":"8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb"} Dec 03 17:40:58 crc kubenswrapper[4787]: I1203 17:40:58.328128 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c5f585a-69da-4588-ab44-bc08513f4c9f","Type":"ContainerStarted","Data":"db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9"} Dec 03 17:41:00 crc kubenswrapper[4787]: I1203 17:41:00.348315 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c5f585a-69da-4588-ab44-bc08513f4c9f","Type":"ContainerStarted","Data":"4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa"} Dec 03 17:41:00 crc kubenswrapper[4787]: I1203 17:41:00.349858 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 17:41:00 crc kubenswrapper[4787]: I1203 17:41:00.376768 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.74405611 podStartE2EDuration="9.376751561s" podCreationTimestamp="2025-12-03 17:40:51 +0000 UTC" firstStartedPulling="2025-12-03 17:40:52.372745502 +0000 UTC m=+1669.190216461" lastFinishedPulling="2025-12-03 17:40:59.005440953 +0000 UTC m=+1675.822911912" observedRunningTime="2025-12-03 17:41:00.371238633 +0000 UTC m=+1677.188709592" watchObservedRunningTime="2025-12-03 17:41:00.376751561 +0000 UTC m=+1677.194222520" Dec 03 17:41:00 crc kubenswrapper[4787]: I1203 17:41:00.856227 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:41:00 crc kubenswrapper[4787]: I1203 17:41:00.927075 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-whzrl"] Dec 03 17:41:00 crc kubenswrapper[4787]: I1203 17:41:00.927343 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" podUID="9ac14488-1fe6-458a-a812-e4b62016f67f" containerName="dnsmasq-dns" containerID="cri-o://2669ec334559ba3f1ee8d39ff64ee6f2897545548e3904e7bb411fee93de4aae" gracePeriod=10 Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.129878 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-hfmfg"] Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.132214 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.149856 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-hfmfg"] Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.193106 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-swift-storage-0\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.193160 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-openstack-edpm-ipam\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.193187 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-svc\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.193204 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8hgb\" (UniqueName: \"kubernetes.io/projected/7e131b5d-ae30-4381-b9ec-b14feffb63ee-kube-api-access-s8hgb\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.193433 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-config\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.193488 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-nb\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.193540 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-sb\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.295737 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-nb\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.295835 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-sb\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.295901 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-swift-storage-0\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.295937 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-openstack-edpm-ipam\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.295967 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-svc\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.295987 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8hgb\" (UniqueName: \"kubernetes.io/projected/7e131b5d-ae30-4381-b9ec-b14feffb63ee-kube-api-access-s8hgb\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.296099 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-config\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.297203 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-svc\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.297547 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-openstack-edpm-ipam\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.297835 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-nb\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.298254 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-swift-storage-0\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.298776 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-config\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.299167 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-sb\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.320505 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8hgb\" (UniqueName: \"kubernetes.io/projected/7e131b5d-ae30-4381-b9ec-b14feffb63ee-kube-api-access-s8hgb\") pod \"dnsmasq-dns-6559847fc9-hfmfg\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.418348 4787 generic.go:334] "Generic (PLEG): container finished" podID="9ac14488-1fe6-458a-a812-e4b62016f67f" containerID="2669ec334559ba3f1ee8d39ff64ee6f2897545548e3904e7bb411fee93de4aae" exitCode=0 Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.418776 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" event={"ID":"9ac14488-1fe6-458a-a812-e4b62016f67f","Type":"ContainerDied","Data":"2669ec334559ba3f1ee8d39ff64ee6f2897545548e3904e7bb411fee93de4aae"} Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.462717 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.609089 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.713835 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-swift-storage-0\") pod \"9ac14488-1fe6-458a-a812-e4b62016f67f\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.713916 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gtlpz\" (UniqueName: \"kubernetes.io/projected/9ac14488-1fe6-458a-a812-e4b62016f67f-kube-api-access-gtlpz\") pod \"9ac14488-1fe6-458a-a812-e4b62016f67f\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.714009 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-svc\") pod \"9ac14488-1fe6-458a-a812-e4b62016f67f\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.714067 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-nb\") pod \"9ac14488-1fe6-458a-a812-e4b62016f67f\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.714140 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-sb\") pod \"9ac14488-1fe6-458a-a812-e4b62016f67f\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.714189 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-config\") pod \"9ac14488-1fe6-458a-a812-e4b62016f67f\" (UID: \"9ac14488-1fe6-458a-a812-e4b62016f67f\") " Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.760445 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9ac14488-1fe6-458a-a812-e4b62016f67f-kube-api-access-gtlpz" (OuterVolumeSpecName: "kube-api-access-gtlpz") pod "9ac14488-1fe6-458a-a812-e4b62016f67f" (UID: "9ac14488-1fe6-458a-a812-e4b62016f67f"). InnerVolumeSpecName "kube-api-access-gtlpz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.820701 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gtlpz\" (UniqueName: \"kubernetes.io/projected/9ac14488-1fe6-458a-a812-e4b62016f67f-kube-api-access-gtlpz\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.867087 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "9ac14488-1fe6-458a-a812-e4b62016f67f" (UID: "9ac14488-1fe6-458a-a812-e4b62016f67f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.867278 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-config" (OuterVolumeSpecName: "config") pod "9ac14488-1fe6-458a-a812-e4b62016f67f" (UID: "9ac14488-1fe6-458a-a812-e4b62016f67f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.873828 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "9ac14488-1fe6-458a-a812-e4b62016f67f" (UID: "9ac14488-1fe6-458a-a812-e4b62016f67f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.877516 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "9ac14488-1fe6-458a-a812-e4b62016f67f" (UID: "9ac14488-1fe6-458a-a812-e4b62016f67f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.907893 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "9ac14488-1fe6-458a-a812-e4b62016f67f" (UID: "9ac14488-1fe6-458a-a812-e4b62016f67f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.922529 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.922820 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.922910 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.922975 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:01 crc kubenswrapper[4787]: I1203 17:41:01.923048 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/9ac14488-1fe6-458a-a812-e4b62016f67f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:02 crc kubenswrapper[4787]: I1203 17:41:02.058041 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-hfmfg"] Dec 03 17:41:02 crc kubenswrapper[4787]: I1203 17:41:02.430677 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" event={"ID":"7e131b5d-ae30-4381-b9ec-b14feffb63ee","Type":"ContainerStarted","Data":"f8c1c616cbf2ebb0cdde7dabec1e2235a993486bc8f294188f8c9c14ba7b39a0"} Dec 03 17:41:02 crc kubenswrapper[4787]: I1203 17:41:02.430724 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" event={"ID":"7e131b5d-ae30-4381-b9ec-b14feffb63ee","Type":"ContainerStarted","Data":"9db35b5bd5868f9d73eff33c7d83a47df7d2b02b371f2ea2feddc5dbe05e566d"} Dec 03 17:41:02 crc kubenswrapper[4787]: I1203 17:41:02.436709 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" Dec 03 17:41:02 crc kubenswrapper[4787]: I1203 17:41:02.436702 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6b7bbf7cf9-whzrl" event={"ID":"9ac14488-1fe6-458a-a812-e4b62016f67f","Type":"ContainerDied","Data":"6ffd935f48feb5a51bde831cf97ce2cdeed11a4270d24095a1cba5af42864dfc"} Dec 03 17:41:02 crc kubenswrapper[4787]: I1203 17:41:02.436785 4787 scope.go:117] "RemoveContainer" containerID="2669ec334559ba3f1ee8d39ff64ee6f2897545548e3904e7bb411fee93de4aae" Dec 03 17:41:02 crc kubenswrapper[4787]: I1203 17:41:02.477779 4787 scope.go:117] "RemoveContainer" containerID="8ef8ae651004d83911726f1a859674c0eaea06e6d301fb5285a3bbcf04b6e519" Dec 03 17:41:02 crc kubenswrapper[4787]: I1203 17:41:02.490077 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-whzrl"] Dec 03 17:41:02 crc kubenswrapper[4787]: I1203 17:41:02.499825 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6b7bbf7cf9-whzrl"] Dec 03 17:41:03 crc kubenswrapper[4787]: I1203 17:41:03.447888 4787 generic.go:334] "Generic (PLEG): container finished" podID="7e131b5d-ae30-4381-b9ec-b14feffb63ee" containerID="f8c1c616cbf2ebb0cdde7dabec1e2235a993486bc8f294188f8c9c14ba7b39a0" exitCode=0 Dec 03 17:41:03 crc kubenswrapper[4787]: I1203 17:41:03.447986 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" event={"ID":"7e131b5d-ae30-4381-b9ec-b14feffb63ee","Type":"ContainerDied","Data":"f8c1c616cbf2ebb0cdde7dabec1e2235a993486bc8f294188f8c9c14ba7b39a0"} Dec 03 17:41:03 crc kubenswrapper[4787]: I1203 17:41:03.779230 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9ac14488-1fe6-458a-a812-e4b62016f67f" path="/var/lib/kubelet/pods/9ac14488-1fe6-458a-a812-e4b62016f67f/volumes" Dec 03 17:41:04 crc kubenswrapper[4787]: I1203 17:41:04.462293 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" event={"ID":"7e131b5d-ae30-4381-b9ec-b14feffb63ee","Type":"ContainerStarted","Data":"130b047da394f7af9e253c5cd18ca0b26b05e55da8bafe1287b0c171a0fc939a"} Dec 03 17:41:04 crc kubenswrapper[4787]: I1203 17:41:04.462843 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:04 crc kubenswrapper[4787]: I1203 17:41:04.464440 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2gfq8" event={"ID":"5749d651-999d-4d54-8661-1962fc41521c","Type":"ContainerStarted","Data":"beed253452ae9872609aaefd211c3936f663e0f5d08560a30440cbccbe3c7e20"} Dec 03 17:41:04 crc kubenswrapper[4787]: I1203 17:41:04.488004 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" podStartSLOduration=3.487983776 podStartE2EDuration="3.487983776s" podCreationTimestamp="2025-12-03 17:41:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:41:04.480063841 +0000 UTC m=+1681.297534830" watchObservedRunningTime="2025-12-03 17:41:04.487983776 +0000 UTC m=+1681.305454735" Dec 03 17:41:04 crc kubenswrapper[4787]: I1203 17:41:04.507848 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-db-sync-2gfq8" podStartSLOduration=1.5868547149999999 podStartE2EDuration="34.507827364s" podCreationTimestamp="2025-12-03 17:40:30 +0000 UTC" firstStartedPulling="2025-12-03 17:40:31.031185485 +0000 UTC m=+1647.848656454" lastFinishedPulling="2025-12-03 17:41:03.952158144 +0000 UTC m=+1680.769629103" observedRunningTime="2025-12-03 17:41:04.495825519 +0000 UTC m=+1681.313296498" watchObservedRunningTime="2025-12-03 17:41:04.507827364 +0000 UTC m=+1681.325298323" Dec 03 17:41:10 crc kubenswrapper[4787]: I1203 17:41:10.530920 4787 generic.go:334] "Generic (PLEG): container finished" podID="5749d651-999d-4d54-8661-1962fc41521c" containerID="beed253452ae9872609aaefd211c3936f663e0f5d08560a30440cbccbe3c7e20" exitCode=0 Dec 03 17:41:10 crc kubenswrapper[4787]: I1203 17:41:10.530999 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2gfq8" event={"ID":"5749d651-999d-4d54-8661-1962fc41521c","Type":"ContainerDied","Data":"beed253452ae9872609aaefd211c3936f663e0f5d08560a30440cbccbe3c7e20"} Dec 03 17:41:11 crc kubenswrapper[4787]: I1203 17:41:11.465218 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 17:41:11 crc kubenswrapper[4787]: I1203 17:41:11.575207 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-nstbb"] Dec 03 17:41:11 crc kubenswrapper[4787]: I1203 17:41:11.576603 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" podUID="b1660890-8f51-407f-ba9d-c55853e01702" containerName="dnsmasq-dns" containerID="cri-o://3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643" gracePeriod=10 Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.177069 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2gfq8" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.187085 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.260963 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-config\") pod \"b1660890-8f51-407f-ba9d-c55853e01702\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.261185 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-frlth\" (UniqueName: \"kubernetes.io/projected/b1660890-8f51-407f-ba9d-c55853e01702-kube-api-access-frlth\") pod \"b1660890-8f51-407f-ba9d-c55853e01702\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.261364 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-sb\") pod \"b1660890-8f51-407f-ba9d-c55853e01702\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.261452 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-nb\") pod \"b1660890-8f51-407f-ba9d-c55853e01702\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.261562 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-swift-storage-0\") pod \"b1660890-8f51-407f-ba9d-c55853e01702\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.261608 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27wt9\" (UniqueName: \"kubernetes.io/projected/5749d651-999d-4d54-8661-1962fc41521c-kube-api-access-27wt9\") pod \"5749d651-999d-4d54-8661-1962fc41521c\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.261690 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-svc\") pod \"b1660890-8f51-407f-ba9d-c55853e01702\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.261753 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-config-data\") pod \"5749d651-999d-4d54-8661-1962fc41521c\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.261816 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-combined-ca-bundle\") pod \"5749d651-999d-4d54-8661-1962fc41521c\" (UID: \"5749d651-999d-4d54-8661-1962fc41521c\") " Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.261894 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-openstack-edpm-ipam\") pod \"b1660890-8f51-407f-ba9d-c55853e01702\" (UID: \"b1660890-8f51-407f-ba9d-c55853e01702\") " Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.267995 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1660890-8f51-407f-ba9d-c55853e01702-kube-api-access-frlth" (OuterVolumeSpecName: "kube-api-access-frlth") pod "b1660890-8f51-407f-ba9d-c55853e01702" (UID: "b1660890-8f51-407f-ba9d-c55853e01702"). InnerVolumeSpecName "kube-api-access-frlth". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.274402 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5749d651-999d-4d54-8661-1962fc41521c-kube-api-access-27wt9" (OuterVolumeSpecName: "kube-api-access-27wt9") pod "5749d651-999d-4d54-8661-1962fc41521c" (UID: "5749d651-999d-4d54-8661-1962fc41521c"). InnerVolumeSpecName "kube-api-access-27wt9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.295580 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5749d651-999d-4d54-8661-1962fc41521c" (UID: "5749d651-999d-4d54-8661-1962fc41521c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.329104 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b1660890-8f51-407f-ba9d-c55853e01702" (UID: "b1660890-8f51-407f-ba9d-c55853e01702"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.329225 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "b1660890-8f51-407f-ba9d-c55853e01702" (UID: "b1660890-8f51-407f-ba9d-c55853e01702"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.335793 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "b1660890-8f51-407f-ba9d-c55853e01702" (UID: "b1660890-8f51-407f-ba9d-c55853e01702"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.340363 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "b1660890-8f51-407f-ba9d-c55853e01702" (UID: "b1660890-8f51-407f-ba9d-c55853e01702"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.341297 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b1660890-8f51-407f-ba9d-c55853e01702" (UID: "b1660890-8f51-407f-ba9d-c55853e01702"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.345846 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-config" (OuterVolumeSpecName: "config") pod "b1660890-8f51-407f-ba9d-c55853e01702" (UID: "b1660890-8f51-407f-ba9d-c55853e01702"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.364769 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.364819 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27wt9\" (UniqueName: \"kubernetes.io/projected/5749d651-999d-4d54-8661-1962fc41521c-kube-api-access-27wt9\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.364835 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.364847 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.364858 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.364868 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-config\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.364879 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-frlth\" (UniqueName: \"kubernetes.io/projected/b1660890-8f51-407f-ba9d-c55853e01702-kube-api-access-frlth\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.364890 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.364899 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/b1660890-8f51-407f-ba9d-c55853e01702-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.370895 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-config-data" (OuterVolumeSpecName: "config-data") pod "5749d651-999d-4d54-8661-1962fc41521c" (UID: "5749d651-999d-4d54-8661-1962fc41521c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.466956 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5749d651-999d-4d54-8661-1962fc41521c-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.554463 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-db-sync-2gfq8" event={"ID":"5749d651-999d-4d54-8661-1962fc41521c","Type":"ContainerDied","Data":"32882de188c35d334659b6e473f4f394a1c53cf018ef993169059cd100552012"} Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.554837 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32882de188c35d334659b6e473f4f394a1c53cf018ef993169059cd100552012" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.554917 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-db-sync-2gfq8" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.556909 4787 generic.go:334] "Generic (PLEG): container finished" podID="b1660890-8f51-407f-ba9d-c55853e01702" containerID="3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643" exitCode=0 Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.556965 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.556970 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" event={"ID":"b1660890-8f51-407f-ba9d-c55853e01702","Type":"ContainerDied","Data":"3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643"} Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.557060 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7d84b4d45c-nstbb" event={"ID":"b1660890-8f51-407f-ba9d-c55853e01702","Type":"ContainerDied","Data":"178844243ac3037e1fa8448e59f097d84d4acee48fd89bc173968af840dbbd1b"} Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.557082 4787 scope.go:117] "RemoveContainer" containerID="3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.600708 4787 scope.go:117] "RemoveContainer" containerID="4cd58953ea8fe9ca6599b8d1eebbcd8d5d6b6c22b3b4f490e842326d9184327f" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.635875 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-nstbb"] Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.647371 4787 scope.go:117] "RemoveContainer" containerID="3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643" Dec 03 17:41:12 crc kubenswrapper[4787]: E1203 17:41:12.648635 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643\": container with ID starting with 3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643 not found: ID does not exist" containerID="3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.648675 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643"} err="failed to get container status \"3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643\": rpc error: code = NotFound desc = could not find container \"3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643\": container with ID starting with 3f1f428a614f851c8b4df6799ca0dd8288478f56d8b8f3b4f112d93e1d2e5643 not found: ID does not exist" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.648699 4787 scope.go:117] "RemoveContainer" containerID="4cd58953ea8fe9ca6599b8d1eebbcd8d5d6b6c22b3b4f490e842326d9184327f" Dec 03 17:41:12 crc kubenswrapper[4787]: E1203 17:41:12.649098 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cd58953ea8fe9ca6599b8d1eebbcd8d5d6b6c22b3b4f490e842326d9184327f\": container with ID starting with 4cd58953ea8fe9ca6599b8d1eebbcd8d5d6b6c22b3b4f490e842326d9184327f not found: ID does not exist" containerID="4cd58953ea8fe9ca6599b8d1eebbcd8d5d6b6c22b3b4f490e842326d9184327f" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.649146 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cd58953ea8fe9ca6599b8d1eebbcd8d5d6b6c22b3b4f490e842326d9184327f"} err="failed to get container status \"4cd58953ea8fe9ca6599b8d1eebbcd8d5d6b6c22b3b4f490e842326d9184327f\": rpc error: code = NotFound desc = could not find container \"4cd58953ea8fe9ca6599b8d1eebbcd8d5d6b6c22b3b4f490e842326d9184327f\": container with ID starting with 4cd58953ea8fe9ca6599b8d1eebbcd8d5d6b6c22b3b4f490e842326d9184327f not found: ID does not exist" Dec 03 17:41:12 crc kubenswrapper[4787]: I1203 17:41:12.649919 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d84b4d45c-nstbb"] Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.596870 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-engine-57d98bb984-f2bhk"] Dec 03 17:41:13 crc kubenswrapper[4787]: E1203 17:41:13.597780 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5749d651-999d-4d54-8661-1962fc41521c" containerName="heat-db-sync" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.597801 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5749d651-999d-4d54-8661-1962fc41521c" containerName="heat-db-sync" Dec 03 17:41:13 crc kubenswrapper[4787]: E1203 17:41:13.597859 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1660890-8f51-407f-ba9d-c55853e01702" containerName="dnsmasq-dns" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.597869 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1660890-8f51-407f-ba9d-c55853e01702" containerName="dnsmasq-dns" Dec 03 17:41:13 crc kubenswrapper[4787]: E1203 17:41:13.597883 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac14488-1fe6-458a-a812-e4b62016f67f" containerName="dnsmasq-dns" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.597893 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac14488-1fe6-458a-a812-e4b62016f67f" containerName="dnsmasq-dns" Dec 03 17:41:13 crc kubenswrapper[4787]: E1203 17:41:13.597906 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1660890-8f51-407f-ba9d-c55853e01702" containerName="init" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.597913 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1660890-8f51-407f-ba9d-c55853e01702" containerName="init" Dec 03 17:41:13 crc kubenswrapper[4787]: E1203 17:41:13.597924 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9ac14488-1fe6-458a-a812-e4b62016f67f" containerName="init" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.597930 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9ac14488-1fe6-458a-a812-e4b62016f67f" containerName="init" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.598166 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5749d651-999d-4d54-8661-1962fc41521c" containerName="heat-db-sync" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.598195 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9ac14488-1fe6-458a-a812-e4b62016f67f" containerName="dnsmasq-dns" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.598204 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1660890-8f51-407f-ba9d-c55853e01702" containerName="dnsmasq-dns" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.598990 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.607245 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-57d98bb984-f2bhk"] Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.678669 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-api-74b78cc757-dx9cl"] Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.680202 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.691188 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9019347-02c2-4e52-86a8-07da730bb3b1-combined-ca-bundle\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.691308 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9019347-02c2-4e52-86a8-07da730bb3b1-config-data\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.691334 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6jftk\" (UniqueName: \"kubernetes.io/projected/e9019347-02c2-4e52-86a8-07da730bb3b1-kube-api-access-6jftk\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.691398 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e9019347-02c2-4e52-86a8-07da730bb3b1-config-data-custom\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.710149 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-74b78cc757-dx9cl"] Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.739835 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/heat-cfnapi-796cccd499-c2cvd"] Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.741446 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.796419 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-config-data-custom\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.796547 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9019347-02c2-4e52-86a8-07da730bb3b1-config-data\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.796638 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8smxw\" (UniqueName: \"kubernetes.io/projected/2e92220b-62d7-4d80-8cfa-30a84750f99b-kube-api-access-8smxw\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.796686 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6jftk\" (UniqueName: \"kubernetes.io/projected/e9019347-02c2-4e52-86a8-07da730bb3b1-kube-api-access-6jftk\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.796776 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-combined-ca-bundle\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.796876 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-public-tls-certs\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.796987 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-config-data\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.797229 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e9019347-02c2-4e52-86a8-07da730bb3b1-config-data-custom\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.797320 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-config-data-custom\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.797449 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnlqx\" (UniqueName: \"kubernetes.io/projected/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-kube-api-access-gnlqx\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.797492 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-combined-ca-bundle\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.797611 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-internal-tls-certs\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.797739 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-public-tls-certs\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.797803 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-config-data\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.798079 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9019347-02c2-4e52-86a8-07da730bb3b1-combined-ca-bundle\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.798211 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-internal-tls-certs\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.808595 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e9019347-02c2-4e52-86a8-07da730bb3b1-config-data-custom\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.823080 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e9019347-02c2-4e52-86a8-07da730bb3b1-config-data\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.849692 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e9019347-02c2-4e52-86a8-07da730bb3b1-combined-ca-bundle\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.860393 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6jftk\" (UniqueName: \"kubernetes.io/projected/e9019347-02c2-4e52-86a8-07da730bb3b1-kube-api-access-6jftk\") pod \"heat-engine-57d98bb984-f2bhk\" (UID: \"e9019347-02c2-4e52-86a8-07da730bb3b1\") " pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.862825 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1660890-8f51-407f-ba9d-c55853e01702" path="/var/lib/kubelet/pods/b1660890-8f51-407f-ba9d-c55853e01702/volumes" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.863653 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-796cccd499-c2cvd"] Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901171 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-config-data\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901315 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-internal-tls-certs\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901398 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-config-data-custom\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901427 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8smxw\" (UniqueName: \"kubernetes.io/projected/2e92220b-62d7-4d80-8cfa-30a84750f99b-kube-api-access-8smxw\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901455 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-combined-ca-bundle\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901499 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-public-tls-certs\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901523 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-config-data\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901619 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-config-data-custom\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901661 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnlqx\" (UniqueName: \"kubernetes.io/projected/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-kube-api-access-gnlqx\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901688 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-combined-ca-bundle\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901711 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-internal-tls-certs\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.901736 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-public-tls-certs\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.905942 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-config-data-custom\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.907621 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-internal-tls-certs\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.908376 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-combined-ca-bundle\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.909816 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-config-data\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.910040 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-combined-ca-bundle\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.911456 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-public-tls-certs\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.914612 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-public-tls-certs\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.915157 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-config-data\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.920222 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.920684 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-config-data-custom\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.920864 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e92220b-62d7-4d80-8cfa-30a84750f99b-internal-tls-certs\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.923915 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnlqx\" (UniqueName: \"kubernetes.io/projected/ce3bd00e-d1d3-4086-aa4d-d4a48143f533-kube-api-access-gnlqx\") pod \"heat-cfnapi-796cccd499-c2cvd\" (UID: \"ce3bd00e-d1d3-4086-aa4d-d4a48143f533\") " pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:13 crc kubenswrapper[4787]: I1203 17:41:13.931654 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8smxw\" (UniqueName: \"kubernetes.io/projected/2e92220b-62d7-4d80-8cfa-30a84750f99b-kube-api-access-8smxw\") pod \"heat-api-74b78cc757-dx9cl\" (UID: \"2e92220b-62d7-4d80-8cfa-30a84750f99b\") " pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:14 crc kubenswrapper[4787]: I1203 17:41:14.010745 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:14 crc kubenswrapper[4787]: I1203 17:41:14.075997 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:14 crc kubenswrapper[4787]: I1203 17:41:14.443299 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-engine-57d98bb984-f2bhk"] Dec 03 17:41:14 crc kubenswrapper[4787]: W1203 17:41:14.448073 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode9019347_02c2_4e52_86a8_07da730bb3b1.slice/crio-24ed8424100d29cfdde56dd0e39dc06b793b4718de84ad87fdddae25afba5c78 WatchSource:0}: Error finding container 24ed8424100d29cfdde56dd0e39dc06b793b4718de84ad87fdddae25afba5c78: Status 404 returned error can't find the container with id 24ed8424100d29cfdde56dd0e39dc06b793b4718de84ad87fdddae25afba5c78 Dec 03 17:41:14 crc kubenswrapper[4787]: I1203 17:41:14.586812 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-57d98bb984-f2bhk" event={"ID":"e9019347-02c2-4e52-86a8-07da730bb3b1","Type":"ContainerStarted","Data":"24ed8424100d29cfdde56dd0e39dc06b793b4718de84ad87fdddae25afba5c78"} Dec 03 17:41:14 crc kubenswrapper[4787]: I1203 17:41:14.592362 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-api-74b78cc757-dx9cl"] Dec 03 17:41:14 crc kubenswrapper[4787]: I1203 17:41:14.714504 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/heat-cfnapi-796cccd499-c2cvd"] Dec 03 17:41:15 crc kubenswrapper[4787]: I1203 17:41:15.598672 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-796cccd499-c2cvd" event={"ID":"ce3bd00e-d1d3-4086-aa4d-d4a48143f533","Type":"ContainerStarted","Data":"6dffa22129c11e63fe0337bda93f779986e48419bd6f0af0a7b1332fc3427f5d"} Dec 03 17:41:15 crc kubenswrapper[4787]: I1203 17:41:15.601720 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-57d98bb984-f2bhk" event={"ID":"e9019347-02c2-4e52-86a8-07da730bb3b1","Type":"ContainerStarted","Data":"411e89189f63c62eb7ad0525f3a70588e478709eac65a872ed8e4d294bf2a9fe"} Dec 03 17:41:15 crc kubenswrapper[4787]: I1203 17:41:15.601842 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:15 crc kubenswrapper[4787]: I1203 17:41:15.602836 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-74b78cc757-dx9cl" event={"ID":"2e92220b-62d7-4d80-8cfa-30a84750f99b","Type":"ContainerStarted","Data":"7d5b374ef2e2a9af069be9eff1645e8949fd01e3570cd9eadfc2f345ef5f37c9"} Dec 03 17:41:15 crc kubenswrapper[4787]: I1203 17:41:15.662685 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-engine-57d98bb984-f2bhk" podStartSLOduration=2.6626616050000003 podStartE2EDuration="2.662661605s" podCreationTimestamp="2025-12-03 17:41:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:41:15.654377454 +0000 UTC m=+1692.471848413" watchObservedRunningTime="2025-12-03 17:41:15.662661605 +0000 UTC m=+1692.480132564" Dec 03 17:41:17 crc kubenswrapper[4787]: I1203 17:41:17.627539 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-74b78cc757-dx9cl" event={"ID":"2e92220b-62d7-4d80-8cfa-30a84750f99b","Type":"ContainerStarted","Data":"a8dc88969fb99155e941765d1fad56f20fc51f2af9a405204f6a7f719fba3c2f"} Dec 03 17:41:17 crc kubenswrapper[4787]: I1203 17:41:17.629191 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:17 crc kubenswrapper[4787]: I1203 17:41:17.633971 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-796cccd499-c2cvd" event={"ID":"ce3bd00e-d1d3-4086-aa4d-d4a48143f533","Type":"ContainerStarted","Data":"d79d897c6bbcaf7fdf361d81a81d07c4e37ec0f1cc6009aecbc7d8903a4e319d"} Dec 03 17:41:17 crc kubenswrapper[4787]: I1203 17:41:17.634620 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:17 crc kubenswrapper[4787]: I1203 17:41:17.654837 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-api-74b78cc757-dx9cl" podStartSLOduration=2.975704779 podStartE2EDuration="4.654817307s" podCreationTimestamp="2025-12-03 17:41:13 +0000 UTC" firstStartedPulling="2025-12-03 17:41:14.597782455 +0000 UTC m=+1691.415253414" lastFinishedPulling="2025-12-03 17:41:16.276894983 +0000 UTC m=+1693.094365942" observedRunningTime="2025-12-03 17:41:17.650836171 +0000 UTC m=+1694.468307130" watchObservedRunningTime="2025-12-03 17:41:17.654817307 +0000 UTC m=+1694.472288256" Dec 03 17:41:17 crc kubenswrapper[4787]: I1203 17:41:17.688307 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/heat-cfnapi-796cccd499-c2cvd" podStartSLOduration=3.132925132 podStartE2EDuration="4.68828611s" podCreationTimestamp="2025-12-03 17:41:13 +0000 UTC" firstStartedPulling="2025-12-03 17:41:14.723144658 +0000 UTC m=+1691.540615617" lastFinishedPulling="2025-12-03 17:41:16.278505636 +0000 UTC m=+1693.095976595" observedRunningTime="2025-12-03 17:41:17.672440557 +0000 UTC m=+1694.489911526" watchObservedRunningTime="2025-12-03 17:41:17.68828611 +0000 UTC m=+1694.505757079" Dec 03 17:41:17 crc kubenswrapper[4787]: I1203 17:41:17.713250 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="46f18248-295b-4cae-b29e-51f3cd414799" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.250:3000/\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 17:41:18 crc kubenswrapper[4787]: I1203 17:41:18.990084 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:41:18 crc kubenswrapper[4787]: I1203 17:41:18.990437 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:41:18 crc kubenswrapper[4787]: I1203 17:41:18.990498 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:41:18 crc kubenswrapper[4787]: I1203 17:41:18.991474 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:41:18 crc kubenswrapper[4787]: I1203 17:41:18.991529 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" gracePeriod=600 Dec 03 17:41:19 crc kubenswrapper[4787]: E1203 17:41:19.121216 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:41:19 crc kubenswrapper[4787]: I1203 17:41:19.663238 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" exitCode=0 Dec 03 17:41:19 crc kubenswrapper[4787]: I1203 17:41:19.663330 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75"} Dec 03 17:41:19 crc kubenswrapper[4787]: I1203 17:41:19.663795 4787 scope.go:117] "RemoveContainer" containerID="8530351cb9d4d0e6acbfd6734059c845290cade8ecc29a25ad6ed13ae5d5db93" Dec 03 17:41:19 crc kubenswrapper[4787]: I1203 17:41:19.664791 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:41:19 crc kubenswrapper[4787]: E1203 17:41:19.665199 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:41:21 crc kubenswrapper[4787]: I1203 17:41:21.588771 4787 scope.go:117] "RemoveContainer" containerID="18805ad3848e3536a98c60875d07221f74240d37ddcd3b7b0b20b92b27d3d959" Dec 03 17:41:21 crc kubenswrapper[4787]: I1203 17:41:21.667096 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.553312 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56"] Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.555535 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.557258 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.557979 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.558445 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.558477 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.574274 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56"] Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.661053 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.661136 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.661323 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nclhc\" (UniqueName: \"kubernetes.io/projected/f30ff726-5f48-483b-a5e4-7f49050157f6-kube-api-access-nclhc\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.661493 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.763761 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.763900 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.763944 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.763978 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nclhc\" (UniqueName: \"kubernetes.io/projected/f30ff726-5f48-483b-a5e4-7f49050157f6-kube-api-access-nclhc\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.770516 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.782565 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.784324 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.795946 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nclhc\" (UniqueName: \"kubernetes.io/projected/f30ff726-5f48-483b-a5e4-7f49050157f6-kube-api-access-nclhc\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:23 crc kubenswrapper[4787]: I1203 17:41:23.880139 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:24 crc kubenswrapper[4787]: I1203 17:41:24.519074 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56"] Dec 03 17:41:24 crc kubenswrapper[4787]: W1203 17:41:24.536966 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf30ff726_5f48_483b_a5e4_7f49050157f6.slice/crio-d954a046bc56ef24fafba34569aae8873ef5b41dafcf3a6ebad1935f4f06500b WatchSource:0}: Error finding container d954a046bc56ef24fafba34569aae8873ef5b41dafcf3a6ebad1935f4f06500b: Status 404 returned error can't find the container with id d954a046bc56ef24fafba34569aae8873ef5b41dafcf3a6ebad1935f4f06500b Dec 03 17:41:24 crc kubenswrapper[4787]: I1203 17:41:24.730345 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" event={"ID":"f30ff726-5f48-483b-a5e4-7f49050157f6","Type":"ContainerStarted","Data":"d954a046bc56ef24fafba34569aae8873ef5b41dafcf3a6ebad1935f4f06500b"} Dec 03 17:41:25 crc kubenswrapper[4787]: I1203 17:41:25.615177 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-api-74b78cc757-dx9cl" Dec 03 17:41:25 crc kubenswrapper[4787]: I1203 17:41:25.693745 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5c7847c888-455sv"] Dec 03 17:41:25 crc kubenswrapper[4787]: I1203 17:41:25.694035 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-api-5c7847c888-455sv" podUID="21511f7a-7ed3-48ac-aaf9-3580fbc19111" containerName="heat-api" containerID="cri-o://f32e2dabb9c3e581fef8022b3a27a1836dce1b3ed6ade71b41d28a8290c08e06" gracePeriod=60 Dec 03 17:41:25 crc kubenswrapper[4787]: I1203 17:41:25.787116 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-cfnapi-796cccd499-c2cvd" Dec 03 17:41:25 crc kubenswrapper[4787]: I1203 17:41:25.847856 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5566764dd4-qqp6l"] Dec 03 17:41:25 crc kubenswrapper[4787]: I1203 17:41:25.848283 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" podUID="92c6dbac-064e-484d-aeee-65678e15863a" containerName="heat-cfnapi" containerID="cri-o://4add3899182083f4a01561f2e932c2e4d8bf2672805901f959479a1689ab80a5" gracePeriod=60 Dec 03 17:41:26 crc kubenswrapper[4787]: I1203 17:41:26.765893 4787 generic.go:334] "Generic (PLEG): container finished" podID="9b8eef40-19d4-4640-9517-f603e62e646f" containerID="52e28b1084238d8e9e3cb69d554cf7061d54c2b4d3d74e258d7b241ad1274498" exitCode=0 Dec 03 17:41:26 crc kubenswrapper[4787]: I1203 17:41:26.766078 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9b8eef40-19d4-4640-9517-f603e62e646f","Type":"ContainerDied","Data":"52e28b1084238d8e9e3cb69d554cf7061d54c2b4d3d74e258d7b241ad1274498"} Dec 03 17:41:26 crc kubenswrapper[4787]: I1203 17:41:26.784303 4787 generic.go:334] "Generic (PLEG): container finished" podID="d774ea58-6127-497a-8948-104e60bfe29b" containerID="a6d6eaceb90dc23604c321c6f66e5ec95d1e94f8a7599bfec0610dba36784ae9" exitCode=0 Dec 03 17:41:26 crc kubenswrapper[4787]: I1203 17:41:26.784362 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d774ea58-6127-497a-8948-104e60bfe29b","Type":"ContainerDied","Data":"a6d6eaceb90dc23604c321c6f66e5ec95d1e94f8a7599bfec0610dba36784ae9"} Dec 03 17:41:27 crc kubenswrapper[4787]: I1203 17:41:27.804077 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"9b8eef40-19d4-4640-9517-f603e62e646f","Type":"ContainerStarted","Data":"0d72a965ed18ad018147578ad44561a76031ea45d71b1165f1416fe5d2528d61"} Dec 03 17:41:27 crc kubenswrapper[4787]: I1203 17:41:27.807369 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"d774ea58-6127-497a-8948-104e60bfe29b","Type":"ContainerStarted","Data":"3ac7c7b9923c3c2b05e7e177b27067864f31d058cb66b18c239d20f77cd3be88"} Dec 03 17:41:27 crc kubenswrapper[4787]: I1203 17:41:27.807719 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 03 17:41:27 crc kubenswrapper[4787]: I1203 17:41:27.841830 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=36.841810037 podStartE2EDuration="36.841810037s" podCreationTimestamp="2025-12-03 17:40:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:41:27.831667837 +0000 UTC m=+1704.649138816" watchObservedRunningTime="2025-12-03 17:41:27.841810037 +0000 UTC m=+1704.659280996" Dec 03 17:41:27 crc kubenswrapper[4787]: I1203 17:41:27.861938 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=36.861918254 podStartE2EDuration="36.861918254s" podCreationTimestamp="2025-12-03 17:40:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:41:27.856730955 +0000 UTC m=+1704.674201914" watchObservedRunningTime="2025-12-03 17:41:27.861918254 +0000 UTC m=+1704.679389203" Dec 03 17:41:28 crc kubenswrapper[4787]: I1203 17:41:28.840608 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5c7847c888-455sv" podUID="21511f7a-7ed3-48ac-aaf9-3580fbc19111" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.206:8004/healthcheck\": read tcp 10.217.0.2:50182->10.217.0.206:8004: read: connection reset by peer" Dec 03 17:41:29 crc kubenswrapper[4787]: I1203 17:41:29.004033 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" podUID="92c6dbac-064e-484d-aeee-65678e15863a" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.207:8000/healthcheck\": read tcp 10.217.0.2:50240->10.217.0.207:8000: read: connection reset by peer" Dec 03 17:41:29 crc kubenswrapper[4787]: I1203 17:41:29.766464 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:41:29 crc kubenswrapper[4787]: E1203 17:41:29.767087 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:41:30 crc kubenswrapper[4787]: I1203 17:41:30.938204 4787 generic.go:334] "Generic (PLEG): container finished" podID="92c6dbac-064e-484d-aeee-65678e15863a" containerID="4add3899182083f4a01561f2e932c2e4d8bf2672805901f959479a1689ab80a5" exitCode=0 Dec 03 17:41:30 crc kubenswrapper[4787]: I1203 17:41:30.938638 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" event={"ID":"92c6dbac-064e-484d-aeee-65678e15863a","Type":"ContainerDied","Data":"4add3899182083f4a01561f2e932c2e4d8bf2672805901f959479a1689ab80a5"} Dec 03 17:41:30 crc kubenswrapper[4787]: I1203 17:41:30.961262 4787 generic.go:334] "Generic (PLEG): container finished" podID="21511f7a-7ed3-48ac-aaf9-3580fbc19111" containerID="f32e2dabb9c3e581fef8022b3a27a1836dce1b3ed6ade71b41d28a8290c08e06" exitCode=0 Dec 03 17:41:30 crc kubenswrapper[4787]: I1203 17:41:30.961303 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5c7847c888-455sv" event={"ID":"21511f7a-7ed3-48ac-aaf9-3580fbc19111","Type":"ContainerDied","Data":"f32e2dabb9c3e581fef8022b3a27a1836dce1b3ed6ade71b41d28a8290c08e06"} Dec 03 17:41:31 crc kubenswrapper[4787]: I1203 17:41:31.737117 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:41:33 crc kubenswrapper[4787]: I1203 17:41:33.953997 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/heat-engine-57d98bb984-f2bhk" Dec 03 17:41:34 crc kubenswrapper[4787]: I1203 17:41:34.019268 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-8b787c4b7-wwr58"] Dec 03 17:41:34 crc kubenswrapper[4787]: I1203 17:41:34.020412 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/heat-engine-8b787c4b7-wwr58" podUID="fef2cdf6-1a2c-4904-9010-c446ad96c3a7" containerName="heat-engine" containerID="cri-o://437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb" gracePeriod=60 Dec 03 17:41:35 crc kubenswrapper[4787]: E1203 17:41:35.442301 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:41:35 crc kubenswrapper[4787]: E1203 17:41:35.444136 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:41:35 crc kubenswrapper[4787]: E1203 17:41:35.446329 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:41:35 crc kubenswrapper[4787]: E1203 17:41:35.446362 4787 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-8b787c4b7-wwr58" podUID="fef2cdf6-1a2c-4904-9010-c446ad96c3a7" containerName="heat-engine" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.596735 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.611687 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.714724 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-public-tls-certs\") pod \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.715313 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nfzw\" (UniqueName: \"kubernetes.io/projected/21511f7a-7ed3-48ac-aaf9-3580fbc19111-kube-api-access-2nfzw\") pod \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.715407 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data\") pod \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.715450 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-public-tls-certs\") pod \"92c6dbac-064e-484d-aeee-65678e15863a\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.715740 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-internal-tls-certs\") pod \"92c6dbac-064e-484d-aeee-65678e15863a\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.716081 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-internal-tls-certs\") pod \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.716158 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data\") pod \"92c6dbac-064e-484d-aeee-65678e15863a\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.716188 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-combined-ca-bundle\") pod \"92c6dbac-064e-484d-aeee-65678e15863a\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.716224 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-combined-ca-bundle\") pod \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.716252 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data-custom\") pod \"92c6dbac-064e-484d-aeee-65678e15863a\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.716472 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hr8l9\" (UniqueName: \"kubernetes.io/projected/92c6dbac-064e-484d-aeee-65678e15863a-kube-api-access-hr8l9\") pod \"92c6dbac-064e-484d-aeee-65678e15863a\" (UID: \"92c6dbac-064e-484d-aeee-65678e15863a\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.716775 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data-custom\") pod \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\" (UID: \"21511f7a-7ed3-48ac-aaf9-3580fbc19111\") " Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.725429 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "92c6dbac-064e-484d-aeee-65678e15863a" (UID: "92c6dbac-064e-484d-aeee-65678e15863a"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.734400 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92c6dbac-064e-484d-aeee-65678e15863a-kube-api-access-hr8l9" (OuterVolumeSpecName: "kube-api-access-hr8l9") pod "92c6dbac-064e-484d-aeee-65678e15863a" (UID: "92c6dbac-064e-484d-aeee-65678e15863a"). InnerVolumeSpecName "kube-api-access-hr8l9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.753647 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "21511f7a-7ed3-48ac-aaf9-3580fbc19111" (UID: "21511f7a-7ed3-48ac-aaf9-3580fbc19111"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.782623 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21511f7a-7ed3-48ac-aaf9-3580fbc19111-kube-api-access-2nfzw" (OuterVolumeSpecName: "kube-api-access-2nfzw") pod "21511f7a-7ed3-48ac-aaf9-3580fbc19111" (UID: "21511f7a-7ed3-48ac-aaf9-3580fbc19111"). InnerVolumeSpecName "kube-api-access-2nfzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.819757 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.819794 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hr8l9\" (UniqueName: \"kubernetes.io/projected/92c6dbac-064e-484d-aeee-65678e15863a-kube-api-access-hr8l9\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.819809 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.819862 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nfzw\" (UniqueName: \"kubernetes.io/projected/21511f7a-7ed3-48ac-aaf9-3580fbc19111-kube-api-access-2nfzw\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.839419 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "21511f7a-7ed3-48ac-aaf9-3580fbc19111" (UID: "21511f7a-7ed3-48ac-aaf9-3580fbc19111"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.860753 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "21511f7a-7ed3-48ac-aaf9-3580fbc19111" (UID: "21511f7a-7ed3-48ac-aaf9-3580fbc19111"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.867243 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92c6dbac-064e-484d-aeee-65678e15863a" (UID: "92c6dbac-064e-484d-aeee-65678e15863a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.872737 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "21511f7a-7ed3-48ac-aaf9-3580fbc19111" (UID: "21511f7a-7ed3-48ac-aaf9-3580fbc19111"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.875389 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "92c6dbac-064e-484d-aeee-65678e15863a" (UID: "92c6dbac-064e-484d-aeee-65678e15863a"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.891985 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "92c6dbac-064e-484d-aeee-65678e15863a" (UID: "92c6dbac-064e-484d-aeee-65678e15863a"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.921917 4787 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.922376 4787 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.922395 4787 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.922409 4787 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.922422 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.922435 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.922109 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data" (OuterVolumeSpecName: "config-data") pod "92c6dbac-064e-484d-aeee-65678e15863a" (UID: "92c6dbac-064e-484d-aeee-65678e15863a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:37 crc kubenswrapper[4787]: I1203 17:41:37.935428 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data" (OuterVolumeSpecName: "config-data") pod "21511f7a-7ed3-48ac-aaf9-3580fbc19111" (UID: "21511f7a-7ed3-48ac-aaf9-3580fbc19111"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.024440 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21511f7a-7ed3-48ac-aaf9-3580fbc19111-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.024475 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/92c6dbac-064e-484d-aeee-65678e15863a-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.062103 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.062126 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" event={"ID":"92c6dbac-064e-484d-aeee-65678e15863a","Type":"ContainerDied","Data":"da20ced467ef59ed0e26b364b2423ad3276a47445a969a983cc58be02e811b46"} Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.062190 4787 scope.go:117] "RemoveContainer" containerID="4add3899182083f4a01561f2e932c2e4d8bf2672805901f959479a1689ab80a5" Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.064992 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-api-5c7847c888-455sv" event={"ID":"21511f7a-7ed3-48ac-aaf9-3580fbc19111","Type":"ContainerDied","Data":"aaa229a64996c290c603b07fed035883e4972f344c980a87e002f4f808fdd7e5"} Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.065065 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-api-5c7847c888-455sv" Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.101575 4787 scope.go:117] "RemoveContainer" containerID="f32e2dabb9c3e581fef8022b3a27a1836dce1b3ed6ade71b41d28a8290c08e06" Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.104524 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-cfnapi-5566764dd4-qqp6l"] Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.114362 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-cfnapi-5566764dd4-qqp6l"] Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.123562 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-api-5c7847c888-455sv"] Dec 03 17:41:38 crc kubenswrapper[4787]: I1203 17:41:38.133244 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-api-5c7847c888-455sv"] Dec 03 17:41:39 crc kubenswrapper[4787]: I1203 17:41:39.081451 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" event={"ID":"f30ff726-5f48-483b-a5e4-7f49050157f6","Type":"ContainerStarted","Data":"c67efae0d5b9257e3f2d0e69bc7e4a0d411c7e3f666e0e644080c84fcc472d71"} Dec 03 17:41:39 crc kubenswrapper[4787]: I1203 17:41:39.101807 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" podStartSLOduration=3.044857065 podStartE2EDuration="16.101783016s" podCreationTimestamp="2025-12-03 17:41:23 +0000 UTC" firstStartedPulling="2025-12-03 17:41:24.539633385 +0000 UTC m=+1701.357104344" lastFinishedPulling="2025-12-03 17:41:37.596559336 +0000 UTC m=+1714.414030295" observedRunningTime="2025-12-03 17:41:39.094936724 +0000 UTC m=+1715.912407693" watchObservedRunningTime="2025-12-03 17:41:39.101783016 +0000 UTC m=+1715.919253975" Dec 03 17:41:39 crc kubenswrapper[4787]: I1203 17:41:39.784419 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21511f7a-7ed3-48ac-aaf9-3580fbc19111" path="/var/lib/kubelet/pods/21511f7a-7ed3-48ac-aaf9-3580fbc19111/volumes" Dec 03 17:41:39 crc kubenswrapper[4787]: I1203 17:41:39.785182 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92c6dbac-064e-484d-aeee-65678e15863a" path="/var/lib/kubelet/pods/92c6dbac-064e-484d-aeee-65678e15863a/volumes" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.289058 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-8jchz"] Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.326974 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-8jchz"] Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.387097 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-db-sync-62rbm"] Dec 03 17:41:40 crc kubenswrapper[4787]: E1203 17:41:40.387748 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c6dbac-064e-484d-aeee-65678e15863a" containerName="heat-cfnapi" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.387772 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c6dbac-064e-484d-aeee-65678e15863a" containerName="heat-cfnapi" Dec 03 17:41:40 crc kubenswrapper[4787]: E1203 17:41:40.387795 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21511f7a-7ed3-48ac-aaf9-3580fbc19111" containerName="heat-api" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.387803 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="21511f7a-7ed3-48ac-aaf9-3580fbc19111" containerName="heat-api" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.387999 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="21511f7a-7ed3-48ac-aaf9-3580fbc19111" containerName="heat-api" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.388055 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="92c6dbac-064e-484d-aeee-65678e15863a" containerName="heat-cfnapi" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.388831 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.391332 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.396864 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-62rbm"] Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.478171 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljr8z\" (UniqueName: \"kubernetes.io/projected/68706b1c-146b-4c74-a8b0-4c24ff3758da-kube-api-access-ljr8z\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.478233 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-config-data\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.478495 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-scripts\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.478549 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-combined-ca-bundle\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.581378 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-scripts\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.581433 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-combined-ca-bundle\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.581620 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljr8z\" (UniqueName: \"kubernetes.io/projected/68706b1c-146b-4c74-a8b0-4c24ff3758da-kube-api-access-ljr8z\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.581660 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-config-data\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.587805 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-combined-ca-bundle\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.589341 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-scripts\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.590450 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-config-data\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.600189 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljr8z\" (UniqueName: \"kubernetes.io/projected/68706b1c-146b-4c74-a8b0-4c24ff3758da-kube-api-access-ljr8z\") pod \"aodh-db-sync-62rbm\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:40 crc kubenswrapper[4787]: I1203 17:41:40.706479 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:41 crc kubenswrapper[4787]: I1203 17:41:41.245406 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-db-sync-62rbm"] Dec 03 17:41:41 crc kubenswrapper[4787]: I1203 17:41:41.741623 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 03 17:41:41 crc kubenswrapper[4787]: I1203 17:41:41.783711 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92bdbf6e-87dc-4a93-aefd-26b002566663" path="/var/lib/kubelet/pods/92bdbf6e-87dc-4a93-aefd-26b002566663/volumes" Dec 03 17:41:41 crc kubenswrapper[4787]: I1203 17:41:41.784680 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 03 17:41:42 crc kubenswrapper[4787]: I1203 17:41:42.119042 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-62rbm" event={"ID":"68706b1c-146b-4c74-a8b0-4c24ff3758da","Type":"ContainerStarted","Data":"ee0ca3330154295df64caaf4d6d3fcbc532c931c6be42ab98983356273dc38a4"} Dec 03 17:41:43 crc kubenswrapper[4787]: I1203 17:41:43.246681 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-api-5c7847c888-455sv" podUID="21511f7a-7ed3-48ac-aaf9-3580fbc19111" containerName="heat-api" probeResult="failure" output="Get \"https://10.217.0.206:8004/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 17:41:43 crc kubenswrapper[4787]: I1203 17:41:43.267589 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/heat-cfnapi-5566764dd4-qqp6l" podUID="92c6dbac-064e-484d-aeee-65678e15863a" containerName="heat-cfnapi" probeResult="failure" output="Get \"https://10.217.0.207:8000/healthcheck\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 03 17:41:43 crc kubenswrapper[4787]: I1203 17:41:43.775222 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:41:43 crc kubenswrapper[4787]: E1203 17:41:43.775840 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:41:45 crc kubenswrapper[4787]: E1203 17:41:45.442001 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:41:45 crc kubenswrapper[4787]: E1203 17:41:45.443867 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:41:45 crc kubenswrapper[4787]: E1203 17:41:45.447662 4787 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" containerID="437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb" cmd=["/usr/bin/pgrep","-r","DRST","heat-engine"] Dec 03 17:41:45 crc kubenswrapper[4787]: E1203 17:41:45.447718 4787 prober.go:104] "Probe errored" err="rpc error: code = Unknown desc = command error: cannot register an exec PID: container is stopping, stdout: , stderr: , exit code -1" probeType="Readiness" pod="openstack/heat-engine-8b787c4b7-wwr58" podUID="fef2cdf6-1a2c-4904-9010-c446ad96c3a7" containerName="heat-engine" Dec 03 17:41:49 crc kubenswrapper[4787]: I1203 17:41:49.214994 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-62rbm" event={"ID":"68706b1c-146b-4c74-a8b0-4c24ff3758da","Type":"ContainerStarted","Data":"ab006eeba764bb3e104337c95c5ba176b4a6b3ab5a4a144b7be08b5338f6b9dc"} Dec 03 17:41:49 crc kubenswrapper[4787]: I1203 17:41:49.238291 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-db-sync-62rbm" podStartSLOduration=1.936333722 podStartE2EDuration="9.23827328s" podCreationTimestamp="2025-12-03 17:41:40 +0000 UTC" firstStartedPulling="2025-12-03 17:41:41.266772088 +0000 UTC m=+1718.084243047" lastFinishedPulling="2025-12-03 17:41:48.568711646 +0000 UTC m=+1725.386182605" observedRunningTime="2025-12-03 17:41:49.232412614 +0000 UTC m=+1726.049883583" watchObservedRunningTime="2025-12-03 17:41:49.23827328 +0000 UTC m=+1726.055744239" Dec 03 17:41:49 crc kubenswrapper[4787]: I1203 17:41:49.990060 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.119419 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-combined-ca-bundle\") pod \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.119529 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xlkpj\" (UniqueName: \"kubernetes.io/projected/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-kube-api-access-xlkpj\") pod \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.119617 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data\") pod \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.119727 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data-custom\") pod \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\" (UID: \"fef2cdf6-1a2c-4904-9010-c446ad96c3a7\") " Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.125567 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-kube-api-access-xlkpj" (OuterVolumeSpecName: "kube-api-access-xlkpj") pod "fef2cdf6-1a2c-4904-9010-c446ad96c3a7" (UID: "fef2cdf6-1a2c-4904-9010-c446ad96c3a7"). InnerVolumeSpecName "kube-api-access-xlkpj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.145814 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "fef2cdf6-1a2c-4904-9010-c446ad96c3a7" (UID: "fef2cdf6-1a2c-4904-9010-c446ad96c3a7"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.161577 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "fef2cdf6-1a2c-4904-9010-c446ad96c3a7" (UID: "fef2cdf6-1a2c-4904-9010-c446ad96c3a7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.177791 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data" (OuterVolumeSpecName: "config-data") pod "fef2cdf6-1a2c-4904-9010-c446ad96c3a7" (UID: "fef2cdf6-1a2c-4904-9010-c446ad96c3a7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.224355 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xlkpj\" (UniqueName: \"kubernetes.io/projected/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-kube-api-access-xlkpj\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.224390 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.224403 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.224416 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fef2cdf6-1a2c-4904-9010-c446ad96c3a7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.227579 4787 generic.go:334] "Generic (PLEG): container finished" podID="f30ff726-5f48-483b-a5e4-7f49050157f6" containerID="c67efae0d5b9257e3f2d0e69bc7e4a0d411c7e3f666e0e644080c84fcc472d71" exitCode=0 Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.227645 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" event={"ID":"f30ff726-5f48-483b-a5e4-7f49050157f6","Type":"ContainerDied","Data":"c67efae0d5b9257e3f2d0e69bc7e4a0d411c7e3f666e0e644080c84fcc472d71"} Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.230923 4787 generic.go:334] "Generic (PLEG): container finished" podID="fef2cdf6-1a2c-4904-9010-c446ad96c3a7" containerID="437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb" exitCode=0 Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.231001 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/heat-engine-8b787c4b7-wwr58" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.231046 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-8b787c4b7-wwr58" event={"ID":"fef2cdf6-1a2c-4904-9010-c446ad96c3a7","Type":"ContainerDied","Data":"437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb"} Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.231116 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/heat-engine-8b787c4b7-wwr58" event={"ID":"fef2cdf6-1a2c-4904-9010-c446ad96c3a7","Type":"ContainerDied","Data":"0817b80698ab5b085607d874587c272e877277250405a26ed23b7f1ce993348a"} Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.231179 4787 scope.go:117] "RemoveContainer" containerID="437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.264211 4787 scope.go:117] "RemoveContainer" containerID="437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb" Dec 03 17:41:50 crc kubenswrapper[4787]: E1203 17:41:50.264755 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb\": container with ID starting with 437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb not found: ID does not exist" containerID="437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.264787 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb"} err="failed to get container status \"437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb\": rpc error: code = NotFound desc = could not find container \"437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb\": container with ID starting with 437963ba4bc3990dbc1f2bc9988bee9bf9414437d3b2ad395ccb1309278311eb not found: ID does not exist" Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.299258 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-engine-8b787c4b7-wwr58"] Dec 03 17:41:50 crc kubenswrapper[4787]: I1203 17:41:50.308629 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-engine-8b787c4b7-wwr58"] Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.735806 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.756963 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nclhc\" (UniqueName: \"kubernetes.io/projected/f30ff726-5f48-483b-a5e4-7f49050157f6-kube-api-access-nclhc\") pod \"f30ff726-5f48-483b-a5e4-7f49050157f6\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.757272 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-repo-setup-combined-ca-bundle\") pod \"f30ff726-5f48-483b-a5e4-7f49050157f6\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.757486 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-inventory\") pod \"f30ff726-5f48-483b-a5e4-7f49050157f6\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.757541 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-ssh-key\") pod \"f30ff726-5f48-483b-a5e4-7f49050157f6\" (UID: \"f30ff726-5f48-483b-a5e4-7f49050157f6\") " Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.762903 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f30ff726-5f48-483b-a5e4-7f49050157f6-kube-api-access-nclhc" (OuterVolumeSpecName: "kube-api-access-nclhc") pod "f30ff726-5f48-483b-a5e4-7f49050157f6" (UID: "f30ff726-5f48-483b-a5e4-7f49050157f6"). InnerVolumeSpecName "kube-api-access-nclhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.766709 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f30ff726-5f48-483b-a5e4-7f49050157f6" (UID: "f30ff726-5f48-483b-a5e4-7f49050157f6"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.785199 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fef2cdf6-1a2c-4904-9010-c446ad96c3a7" path="/var/lib/kubelet/pods/fef2cdf6-1a2c-4904-9010-c446ad96c3a7/volumes" Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.790549 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f30ff726-5f48-483b-a5e4-7f49050157f6" (UID: "f30ff726-5f48-483b-a5e4-7f49050157f6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.803840 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-inventory" (OuterVolumeSpecName: "inventory") pod "f30ff726-5f48-483b-a5e4-7f49050157f6" (UID: "f30ff726-5f48-483b-a5e4-7f49050157f6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.860449 4787 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.860626 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.860685 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f30ff726-5f48-483b-a5e4-7f49050157f6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:51 crc kubenswrapper[4787]: I1203 17:41:51.860777 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nclhc\" (UniqueName: \"kubernetes.io/projected/f30ff726-5f48-483b-a5e4-7f49050157f6-kube-api-access-nclhc\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.252524 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" event={"ID":"f30ff726-5f48-483b-a5e4-7f49050157f6","Type":"ContainerDied","Data":"d954a046bc56ef24fafba34569aae8873ef5b41dafcf3a6ebad1935f4f06500b"} Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.252831 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d954a046bc56ef24fafba34569aae8873ef5b41dafcf3a6ebad1935f4f06500b" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.252546 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.254805 4787 generic.go:334] "Generic (PLEG): container finished" podID="68706b1c-146b-4c74-a8b0-4c24ff3758da" containerID="ab006eeba764bb3e104337c95c5ba176b4a6b3ab5a4a144b7be08b5338f6b9dc" exitCode=0 Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.254830 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-62rbm" event={"ID":"68706b1c-146b-4c74-a8b0-4c24ff3758da","Type":"ContainerDied","Data":"ab006eeba764bb3e104337c95c5ba176b4a6b3ab5a4a144b7be08b5338f6b9dc"} Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.342742 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2"] Dec 03 17:41:52 crc kubenswrapper[4787]: E1203 17:41:52.343232 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f30ff726-5f48-483b-a5e4-7f49050157f6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.343251 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f30ff726-5f48-483b-a5e4-7f49050157f6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 17:41:52 crc kubenswrapper[4787]: E1203 17:41:52.343269 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fef2cdf6-1a2c-4904-9010-c446ad96c3a7" containerName="heat-engine" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.343275 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="fef2cdf6-1a2c-4904-9010-c446ad96c3a7" containerName="heat-engine" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.343509 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f30ff726-5f48-483b-a5e4-7f49050157f6" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.343532 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="fef2cdf6-1a2c-4904-9010-c446ad96c3a7" containerName="heat-engine" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.344265 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.345800 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.346711 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.346827 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.346970 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.353252 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2"] Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.371370 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.371713 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.371915 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.371992 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgjjq\" (UniqueName: \"kubernetes.io/projected/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-kube-api-access-bgjjq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.474488 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.475370 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.476013 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.476228 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgjjq\" (UniqueName: \"kubernetes.io/projected/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-kube-api-access-bgjjq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.479736 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.479760 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.480182 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.496667 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgjjq\" (UniqueName: \"kubernetes.io/projected/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-kube-api-access-bgjjq\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:52 crc kubenswrapper[4787]: I1203 17:41:52.663118 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:41:53 crc kubenswrapper[4787]: W1203 17:41:53.187984 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99e1ef0d_6d9c_413a_9d70_e8b0aa3f441b.slice/crio-56781e71c20a1bb81255d0a686d9c83dff6998cba5b2e5aaf2e82e8ea7f85c6e WatchSource:0}: Error finding container 56781e71c20a1bb81255d0a686d9c83dff6998cba5b2e5aaf2e82e8ea7f85c6e: Status 404 returned error can't find the container with id 56781e71c20a1bb81255d0a686d9c83dff6998cba5b2e5aaf2e82e8ea7f85c6e Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.191995 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2"] Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.265611 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" event={"ID":"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b","Type":"ContainerStarted","Data":"56781e71c20a1bb81255d0a686d9c83dff6998cba5b2e5aaf2e82e8ea7f85c6e"} Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.592343 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.699742 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-combined-ca-bundle\") pod \"68706b1c-146b-4c74-a8b0-4c24ff3758da\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.699860 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-scripts\") pod \"68706b1c-146b-4c74-a8b0-4c24ff3758da\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.700000 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-config-data\") pod \"68706b1c-146b-4c74-a8b0-4c24ff3758da\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.700170 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ljr8z\" (UniqueName: \"kubernetes.io/projected/68706b1c-146b-4c74-a8b0-4c24ff3758da-kube-api-access-ljr8z\") pod \"68706b1c-146b-4c74-a8b0-4c24ff3758da\" (UID: \"68706b1c-146b-4c74-a8b0-4c24ff3758da\") " Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.703922 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-scripts" (OuterVolumeSpecName: "scripts") pod "68706b1c-146b-4c74-a8b0-4c24ff3758da" (UID: "68706b1c-146b-4c74-a8b0-4c24ff3758da"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.704357 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68706b1c-146b-4c74-a8b0-4c24ff3758da-kube-api-access-ljr8z" (OuterVolumeSpecName: "kube-api-access-ljr8z") pod "68706b1c-146b-4c74-a8b0-4c24ff3758da" (UID: "68706b1c-146b-4c74-a8b0-4c24ff3758da"). InnerVolumeSpecName "kube-api-access-ljr8z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.732215 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-config-data" (OuterVolumeSpecName: "config-data") pod "68706b1c-146b-4c74-a8b0-4c24ff3758da" (UID: "68706b1c-146b-4c74-a8b0-4c24ff3758da"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.737272 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "68706b1c-146b-4c74-a8b0-4c24ff3758da" (UID: "68706b1c-146b-4c74-a8b0-4c24ff3758da"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.803714 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.803770 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.803782 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ljr8z\" (UniqueName: \"kubernetes.io/projected/68706b1c-146b-4c74-a8b0-4c24ff3758da-kube-api-access-ljr8z\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:53 crc kubenswrapper[4787]: I1203 17:41:53.803792 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/68706b1c-146b-4c74-a8b0-4c24ff3758da-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:41:54 crc kubenswrapper[4787]: I1203 17:41:54.279036 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" event={"ID":"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b","Type":"ContainerStarted","Data":"7bf32c081a4b702652e0a6f9b5bc7ee04153f83fd34a4535b91f6ade01959fec"} Dec 03 17:41:54 crc kubenswrapper[4787]: I1203 17:41:54.285001 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-db-sync-62rbm" event={"ID":"68706b1c-146b-4c74-a8b0-4c24ff3758da","Type":"ContainerDied","Data":"ee0ca3330154295df64caaf4d6d3fcbc532c931c6be42ab98983356273dc38a4"} Dec 03 17:41:54 crc kubenswrapper[4787]: I1203 17:41:54.285070 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee0ca3330154295df64caaf4d6d3fcbc532c931c6be42ab98983356273dc38a4" Dec 03 17:41:54 crc kubenswrapper[4787]: I1203 17:41:54.285422 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-db-sync-62rbm" Dec 03 17:41:54 crc kubenswrapper[4787]: I1203 17:41:54.326704 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" podStartSLOduration=1.930332624 podStartE2EDuration="2.326685309s" podCreationTimestamp="2025-12-03 17:41:52 +0000 UTC" firstStartedPulling="2025-12-03 17:41:53.190658581 +0000 UTC m=+1730.008129540" lastFinishedPulling="2025-12-03 17:41:53.587011266 +0000 UTC m=+1730.404482225" observedRunningTime="2025-12-03 17:41:54.303070109 +0000 UTC m=+1731.120541068" watchObservedRunningTime="2025-12-03 17:41:54.326685309 +0000 UTC m=+1731.144156288" Dec 03 17:41:55 crc kubenswrapper[4787]: I1203 17:41:55.440656 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 17:41:55 crc kubenswrapper[4787]: I1203 17:41:55.441289 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-api" containerID="cri-o://58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02" gracePeriod=30 Dec 03 17:41:55 crc kubenswrapper[4787]: I1203 17:41:55.441413 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-listener" containerID="cri-o://c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea" gracePeriod=30 Dec 03 17:41:55 crc kubenswrapper[4787]: I1203 17:41:55.441453 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-evaluator" containerID="cri-o://133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5" gracePeriod=30 Dec 03 17:41:55 crc kubenswrapper[4787]: I1203 17:41:55.441531 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/aodh-0" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-notifier" containerID="cri-o://2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67" gracePeriod=30 Dec 03 17:41:56 crc kubenswrapper[4787]: I1203 17:41:56.310213 4787 generic.go:334] "Generic (PLEG): container finished" podID="b51192bf-fe51-4594-ae63-eba136c23535" containerID="133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5" exitCode=0 Dec 03 17:41:56 crc kubenswrapper[4787]: I1203 17:41:56.310505 4787 generic.go:334] "Generic (PLEG): container finished" podID="b51192bf-fe51-4594-ae63-eba136c23535" containerID="58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02" exitCode=0 Dec 03 17:41:56 crc kubenswrapper[4787]: I1203 17:41:56.310307 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b51192bf-fe51-4594-ae63-eba136c23535","Type":"ContainerDied","Data":"133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5"} Dec 03 17:41:56 crc kubenswrapper[4787]: I1203 17:41:56.310542 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b51192bf-fe51-4594-ae63-eba136c23535","Type":"ContainerDied","Data":"58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02"} Dec 03 17:41:58 crc kubenswrapper[4787]: I1203 17:41:58.766670 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:41:58 crc kubenswrapper[4787]: E1203 17:41:58.767454 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:41:59 crc kubenswrapper[4787]: I1203 17:41:59.341851 4787 generic.go:334] "Generic (PLEG): container finished" podID="b51192bf-fe51-4594-ae63-eba136c23535" containerID="c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea" exitCode=0 Dec 03 17:41:59 crc kubenswrapper[4787]: I1203 17:41:59.341952 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b51192bf-fe51-4594-ae63-eba136c23535","Type":"ContainerDied","Data":"c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea"} Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.347642 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.356482 4787 generic.go:334] "Generic (PLEG): container finished" podID="b51192bf-fe51-4594-ae63-eba136c23535" containerID="2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67" exitCode=0 Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.356522 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b51192bf-fe51-4594-ae63-eba136c23535","Type":"ContainerDied","Data":"2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67"} Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.356550 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"b51192bf-fe51-4594-ae63-eba136c23535","Type":"ContainerDied","Data":"70464ed43c1dfd6ac998f0aa03a617d503a935134442b9b4eb750762b885bcb9"} Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.356567 4787 scope.go:117] "RemoveContainer" containerID="c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.356638 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.394762 4787 scope.go:117] "RemoveContainer" containerID="2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.421367 4787 scope.go:117] "RemoveContainer" containerID="133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.448884 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-scripts\") pod \"b51192bf-fe51-4594-ae63-eba136c23535\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.449097 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phjxc\" (UniqueName: \"kubernetes.io/projected/b51192bf-fe51-4594-ae63-eba136c23535-kube-api-access-phjxc\") pod \"b51192bf-fe51-4594-ae63-eba136c23535\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.449120 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-internal-tls-certs\") pod \"b51192bf-fe51-4594-ae63-eba136c23535\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.449245 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-public-tls-certs\") pod \"b51192bf-fe51-4594-ae63-eba136c23535\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.449280 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-combined-ca-bundle\") pod \"b51192bf-fe51-4594-ae63-eba136c23535\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.449347 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-config-data\") pod \"b51192bf-fe51-4594-ae63-eba136c23535\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.456680 4787 scope.go:117] "RemoveContainer" containerID="58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.456997 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b51192bf-fe51-4594-ae63-eba136c23535-kube-api-access-phjxc" (OuterVolumeSpecName: "kube-api-access-phjxc") pod "b51192bf-fe51-4594-ae63-eba136c23535" (UID: "b51192bf-fe51-4594-ae63-eba136c23535"). InnerVolumeSpecName "kube-api-access-phjxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.470603 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-scripts" (OuterVolumeSpecName: "scripts") pod "b51192bf-fe51-4594-ae63-eba136c23535" (UID: "b51192bf-fe51-4594-ae63-eba136c23535"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.544161 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "b51192bf-fe51-4594-ae63-eba136c23535" (UID: "b51192bf-fe51-4594-ae63-eba136c23535"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.550918 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b51192bf-fe51-4594-ae63-eba136c23535" (UID: "b51192bf-fe51-4594-ae63-eba136c23535"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.551129 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-public-tls-certs\") pod \"b51192bf-fe51-4594-ae63-eba136c23535\" (UID: \"b51192bf-fe51-4594-ae63-eba136c23535\") " Dec 03 17:42:00 crc kubenswrapper[4787]: W1203 17:42:00.551335 4787 empty_dir.go:500] Warning: Unmount skipped because path does not exist: /var/lib/kubelet/pods/b51192bf-fe51-4594-ae63-eba136c23535/volumes/kubernetes.io~secret/public-tls-certs Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.551359 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "b51192bf-fe51-4594-ae63-eba136c23535" (UID: "b51192bf-fe51-4594-ae63-eba136c23535"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.551933 4787 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.551955 4787 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.551966 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.551978 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phjxc\" (UniqueName: \"kubernetes.io/projected/b51192bf-fe51-4594-ae63-eba136c23535-kube-api-access-phjxc\") on node \"crc\" DevicePath \"\"" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.612254 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-config-data" (OuterVolumeSpecName: "config-data") pod "b51192bf-fe51-4594-ae63-eba136c23535" (UID: "b51192bf-fe51-4594-ae63-eba136c23535"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.630913 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b51192bf-fe51-4594-ae63-eba136c23535" (UID: "b51192bf-fe51-4594-ae63-eba136c23535"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.655154 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.655206 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b51192bf-fe51-4594-ae63-eba136c23535-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.762930 4787 scope.go:117] "RemoveContainer" containerID="c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea" Dec 03 17:42:00 crc kubenswrapper[4787]: E1203 17:42:00.763461 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea\": container with ID starting with c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea not found: ID does not exist" containerID="c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.763489 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea"} err="failed to get container status \"c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea\": rpc error: code = NotFound desc = could not find container \"c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea\": container with ID starting with c14f76ad1a1d247aa37100bd1cc526c41e39839ba6dd6287a14d927257f460ea not found: ID does not exist" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.763511 4787 scope.go:117] "RemoveContainer" containerID="2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67" Dec 03 17:42:00 crc kubenswrapper[4787]: E1203 17:42:00.763976 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67\": container with ID starting with 2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67 not found: ID does not exist" containerID="2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.764044 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67"} err="failed to get container status \"2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67\": rpc error: code = NotFound desc = could not find container \"2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67\": container with ID starting with 2a32c73955002cc11c1060d479cf4875b45bb4e398c73cc19670607d9d33ac67 not found: ID does not exist" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.764076 4787 scope.go:117] "RemoveContainer" containerID="133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5" Dec 03 17:42:00 crc kubenswrapper[4787]: E1203 17:42:00.764614 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5\": container with ID starting with 133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5 not found: ID does not exist" containerID="133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.764685 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5"} err="failed to get container status \"133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5\": rpc error: code = NotFound desc = could not find container \"133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5\": container with ID starting with 133ca5344524c5a99118d74e88703ca369d08419bac4c3678f1a9337b89f99e5 not found: ID does not exist" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.764710 4787 scope.go:117] "RemoveContainer" containerID="58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02" Dec 03 17:42:00 crc kubenswrapper[4787]: E1203 17:42:00.764995 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02\": container with ID starting with 58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02 not found: ID does not exist" containerID="58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.765034 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02"} err="failed to get container status \"58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02\": rpc error: code = NotFound desc = could not find container \"58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02\": container with ID starting with 58c27ff631a746ef869107a074d65e0910ff388738a79852b0ea2dd358169d02 not found: ID does not exist" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.770719 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0"] Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.789739 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0"] Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.806458 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/aodh-0"] Dec 03 17:42:00 crc kubenswrapper[4787]: E1203 17:42:00.807329 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-api" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.807356 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-api" Dec 03 17:42:00 crc kubenswrapper[4787]: E1203 17:42:00.807391 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-notifier" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.807402 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-notifier" Dec 03 17:42:00 crc kubenswrapper[4787]: E1203 17:42:00.807421 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68706b1c-146b-4c74-a8b0-4c24ff3758da" containerName="aodh-db-sync" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.807432 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="68706b1c-146b-4c74-a8b0-4c24ff3758da" containerName="aodh-db-sync" Dec 03 17:42:00 crc kubenswrapper[4787]: E1203 17:42:00.807484 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-listener" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.807497 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-listener" Dec 03 17:42:00 crc kubenswrapper[4787]: E1203 17:42:00.807515 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-evaluator" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.807525 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-evaluator" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.807780 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-api" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.807808 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-notifier" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.807819 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-listener" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.807831 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="68706b1c-146b-4c74-a8b0-4c24ff3758da" containerName="aodh-db-sync" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.807841 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b51192bf-fe51-4594-ae63-eba136c23535" containerName="aodh-evaluator" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.810754 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.815686 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-internal-svc" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.816081 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-autoscaling-dockercfg-q4n2s" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.816186 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-config-data" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.816263 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"aodh-scripts" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.816307 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-aodh-public-svc" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.819626 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.964675 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-scripts\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.964778 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.964966 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-internal-tls-certs\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.965052 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-public-tls-certs\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.965154 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-config-data\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:00 crc kubenswrapper[4787]: I1203 17:42:00.965390 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkwml\" (UniqueName: \"kubernetes.io/projected/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-kube-api-access-tkwml\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.067689 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-scripts\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.067816 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.067886 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-internal-tls-certs\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.067918 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-public-tls-certs\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.067963 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-config-data\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.068085 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkwml\" (UniqueName: \"kubernetes.io/projected/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-kube-api-access-tkwml\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.076690 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-config-data\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.078706 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-internal-tls-certs\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.079643 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-public-tls-certs\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.081822 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-combined-ca-bundle\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.088540 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-scripts\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.098855 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkwml\" (UniqueName: \"kubernetes.io/projected/33d3a3f3-e9c2-42f3-a13c-34be4404eb1a-kube-api-access-tkwml\") pod \"aodh-0\" (UID: \"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a\") " pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.146666 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/aodh-0" Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.625984 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/aodh-0"] Dec 03 17:42:01 crc kubenswrapper[4787]: I1203 17:42:01.783553 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b51192bf-fe51-4594-ae63-eba136c23535" path="/var/lib/kubelet/pods/b51192bf-fe51-4594-ae63-eba136c23535/volumes" Dec 03 17:42:02 crc kubenswrapper[4787]: I1203 17:42:02.386246 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a","Type":"ContainerStarted","Data":"554b37fa8a976307cafc76a6deed49261da13f8e8e3382015d69d25207c55f91"} Dec 03 17:42:03 crc kubenswrapper[4787]: I1203 17:42:03.399629 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a","Type":"ContainerStarted","Data":"fe447ca2349f898e0b67cb26db964c141ee6736156fd429d26534ea7ab106a6b"} Dec 03 17:42:04 crc kubenswrapper[4787]: I1203 17:42:04.472424 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a","Type":"ContainerStarted","Data":"b397c865753647cd1b5f8e1c4302786f1faf63f22a4eb6b303ffb405a38879db"} Dec 03 17:42:05 crc kubenswrapper[4787]: I1203 17:42:05.488514 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a","Type":"ContainerStarted","Data":"3cb216b15522ff6015ec04532bb16f0c127841fbbde2aa0f83bd61c8e8dfb1e6"} Dec 03 17:42:06 crc kubenswrapper[4787]: I1203 17:42:06.503790 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/aodh-0" event={"ID":"33d3a3f3-e9c2-42f3-a13c-34be4404eb1a","Type":"ContainerStarted","Data":"f8ef23dfe57d234a56de7ff833f7c69292ece6bde77ee98659bebcabd247d17e"} Dec 03 17:42:06 crc kubenswrapper[4787]: I1203 17:42:06.545433 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/aodh-0" podStartSLOduration=2.422543838 podStartE2EDuration="6.545411687s" podCreationTimestamp="2025-12-03 17:42:00 +0000 UTC" firstStartedPulling="2025-12-03 17:42:01.635650134 +0000 UTC m=+1738.453121103" lastFinishedPulling="2025-12-03 17:42:05.758517993 +0000 UTC m=+1742.575988952" observedRunningTime="2025-12-03 17:42:06.529319738 +0000 UTC m=+1743.346790707" watchObservedRunningTime="2025-12-03 17:42:06.545411687 +0000 UTC m=+1743.362882656" Dec 03 17:42:13 crc kubenswrapper[4787]: I1203 17:42:13.799797 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:42:13 crc kubenswrapper[4787]: E1203 17:42:13.800664 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:42:21 crc kubenswrapper[4787]: I1203 17:42:21.817997 4787 scope.go:117] "RemoveContainer" containerID="2e1ab55c40b297261a40fd7c4dfc3e647f8156cad557b8c44cb9f551fb6cfd6f" Dec 03 17:42:21 crc kubenswrapper[4787]: I1203 17:42:21.848878 4787 scope.go:117] "RemoveContainer" containerID="efbc30d5352f0b68539dcfdb6ddff04d64f180b15fccf38745eb784081ca9932" Dec 03 17:42:25 crc kubenswrapper[4787]: I1203 17:42:25.766782 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:42:25 crc kubenswrapper[4787]: E1203 17:42:25.767735 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:42:36 crc kubenswrapper[4787]: I1203 17:42:36.766865 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:42:36 crc kubenswrapper[4787]: E1203 17:42:36.767679 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:42:51 crc kubenswrapper[4787]: I1203 17:42:51.766367 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:42:51 crc kubenswrapper[4787]: E1203 17:42:51.768549 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:43:05 crc kubenswrapper[4787]: I1203 17:43:05.765888 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:43:05 crc kubenswrapper[4787]: E1203 17:43:05.766598 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:43:16 crc kubenswrapper[4787]: I1203 17:43:16.766265 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:43:16 crc kubenswrapper[4787]: E1203 17:43:16.768306 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:43:22 crc kubenswrapper[4787]: I1203 17:43:22.040044 4787 scope.go:117] "RemoveContainer" containerID="df9dd7c608fc2760c54a5d62fabff29c223f72749d743297ee102cb769f0c531" Dec 03 17:43:27 crc kubenswrapper[4787]: I1203 17:43:27.766228 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:43:27 crc kubenswrapper[4787]: E1203 17:43:27.767140 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:43:39 crc kubenswrapper[4787]: I1203 17:43:39.766308 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:43:39 crc kubenswrapper[4787]: E1203 17:43:39.767078 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:43:54 crc kubenswrapper[4787]: I1203 17:43:54.766953 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:43:54 crc kubenswrapper[4787]: E1203 17:43:54.768237 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:44:07 crc kubenswrapper[4787]: I1203 17:44:07.765928 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:44:07 crc kubenswrapper[4787]: E1203 17:44:07.766916 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:44:21 crc kubenswrapper[4787]: I1203 17:44:21.766662 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:44:21 crc kubenswrapper[4787]: E1203 17:44:21.767574 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:44:22 crc kubenswrapper[4787]: I1203 17:44:22.135444 4787 scope.go:117] "RemoveContainer" containerID="93dd278d59121f8b452915cd6cddb7ad9722212416513b86140fcbd6c969f270" Dec 03 17:44:22 crc kubenswrapper[4787]: I1203 17:44:22.203450 4787 scope.go:117] "RemoveContainer" containerID="7d95e709d11a7d7a5fb923a98562238a3910bfc6ece04a7d257edc44eb8008ef" Dec 03 17:44:22 crc kubenswrapper[4787]: I1203 17:44:22.235231 4787 scope.go:117] "RemoveContainer" containerID="7437a6ffa988eb880e4d2bd09c867a8252afe9a4495921f4c650e8d470c59b46" Dec 03 17:44:22 crc kubenswrapper[4787]: I1203 17:44:22.266939 4787 scope.go:117] "RemoveContainer" containerID="866ac7ebfe6c706fe719e1b67dbbe6c90fcc36d8f86913e74ee9d1fe8f2b5f75" Dec 03 17:44:34 crc kubenswrapper[4787]: I1203 17:44:34.766454 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:44:34 crc kubenswrapper[4787]: E1203 17:44:34.767189 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:44:49 crc kubenswrapper[4787]: I1203 17:44:49.767521 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:44:49 crc kubenswrapper[4787]: E1203 17:44:49.769556 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.510502 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hf52w"] Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.514992 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.536546 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hf52w"] Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.565904 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-utilities\") pod \"certified-operators-hf52w\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.566183 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-catalog-content\") pod \"certified-operators-hf52w\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.566237 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swwxc\" (UniqueName: \"kubernetes.io/projected/e854a964-9d6a-4a35-8048-16dfb61fcefc-kube-api-access-swwxc\") pod \"certified-operators-hf52w\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.668186 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-utilities\") pod \"certified-operators-hf52w\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.668310 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-catalog-content\") pod \"certified-operators-hf52w\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.668349 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swwxc\" (UniqueName: \"kubernetes.io/projected/e854a964-9d6a-4a35-8048-16dfb61fcefc-kube-api-access-swwxc\") pod \"certified-operators-hf52w\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.669146 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-utilities\") pod \"certified-operators-hf52w\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.669172 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-catalog-content\") pod \"certified-operators-hf52w\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.694857 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swwxc\" (UniqueName: \"kubernetes.io/projected/e854a964-9d6a-4a35-8048-16dfb61fcefc-kube-api-access-swwxc\") pod \"certified-operators-hf52w\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:44:51 crc kubenswrapper[4787]: I1203 17:44:51.845292 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:44:52 crc kubenswrapper[4787]: I1203 17:44:52.377419 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hf52w"] Dec 03 17:44:52 crc kubenswrapper[4787]: I1203 17:44:52.420833 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hf52w" event={"ID":"e854a964-9d6a-4a35-8048-16dfb61fcefc","Type":"ContainerStarted","Data":"91672280b49c00973f2f9481a196be0bb451f589735c2e9f496d4e5b8573efbb"} Dec 03 17:44:53 crc kubenswrapper[4787]: I1203 17:44:53.434205 4787 generic.go:334] "Generic (PLEG): container finished" podID="e854a964-9d6a-4a35-8048-16dfb61fcefc" containerID="dd23a88e82e2d1f63d43a49dfc968241ba59df7b5608698ab5a83a792aeecd97" exitCode=0 Dec 03 17:44:53 crc kubenswrapper[4787]: I1203 17:44:53.434282 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hf52w" event={"ID":"e854a964-9d6a-4a35-8048-16dfb61fcefc","Type":"ContainerDied","Data":"dd23a88e82e2d1f63d43a49dfc968241ba59df7b5608698ab5a83a792aeecd97"} Dec 03 17:44:54 crc kubenswrapper[4787]: I1203 17:44:54.451844 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hf52w" event={"ID":"e854a964-9d6a-4a35-8048-16dfb61fcefc","Type":"ContainerStarted","Data":"2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810"} Dec 03 17:44:56 crc kubenswrapper[4787]: I1203 17:44:56.479661 4787 generic.go:334] "Generic (PLEG): container finished" podID="e854a964-9d6a-4a35-8048-16dfb61fcefc" containerID="2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810" exitCode=0 Dec 03 17:44:56 crc kubenswrapper[4787]: I1203 17:44:56.479740 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hf52w" event={"ID":"e854a964-9d6a-4a35-8048-16dfb61fcefc","Type":"ContainerDied","Data":"2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810"} Dec 03 17:44:57 crc kubenswrapper[4787]: I1203 17:44:57.497583 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hf52w" event={"ID":"e854a964-9d6a-4a35-8048-16dfb61fcefc","Type":"ContainerStarted","Data":"2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d"} Dec 03 17:44:57 crc kubenswrapper[4787]: I1203 17:44:57.524939 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hf52w" podStartSLOduration=3.072408764 podStartE2EDuration="6.524915128s" podCreationTimestamp="2025-12-03 17:44:51 +0000 UTC" firstStartedPulling="2025-12-03 17:44:53.436701573 +0000 UTC m=+1910.254172542" lastFinishedPulling="2025-12-03 17:44:56.889207947 +0000 UTC m=+1913.706678906" observedRunningTime="2025-12-03 17:44:57.51562533 +0000 UTC m=+1914.333096289" watchObservedRunningTime="2025-12-03 17:44:57.524915128 +0000 UTC m=+1914.342386087" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.052101 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-6bd0-account-create-update-qvtnl"] Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.066320 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-z6tr4"] Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.077202 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8ee8-account-create-update-bvmfg"] Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.086918 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-a415-account-create-update-t8r4c"] Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.097633 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-z6tr4"] Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.135137 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-a415-account-create-update-t8r4c"] Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.158224 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-6bd0-account-create-update-qvtnl"] Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.197336 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8ee8-account-create-update-bvmfg"] Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.240814 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr"] Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.242459 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.245063 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.245545 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.275436 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr"] Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.386081 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/512a131d-22af-40e4-b9ca-a8c04ff844de-config-volume\") pod \"collect-profiles-29413065-8gbhr\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.386132 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/512a131d-22af-40e4-b9ca-a8c04ff844de-secret-volume\") pod \"collect-profiles-29413065-8gbhr\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.386277 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whhw8\" (UniqueName: \"kubernetes.io/projected/512a131d-22af-40e4-b9ca-a8c04ff844de-kube-api-access-whhw8\") pod \"collect-profiles-29413065-8gbhr\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.487643 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whhw8\" (UniqueName: \"kubernetes.io/projected/512a131d-22af-40e4-b9ca-a8c04ff844de-kube-api-access-whhw8\") pod \"collect-profiles-29413065-8gbhr\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.487738 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/512a131d-22af-40e4-b9ca-a8c04ff844de-config-volume\") pod \"collect-profiles-29413065-8gbhr\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.487770 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/512a131d-22af-40e4-b9ca-a8c04ff844de-secret-volume\") pod \"collect-profiles-29413065-8gbhr\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.489439 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/512a131d-22af-40e4-b9ca-a8c04ff844de-config-volume\") pod \"collect-profiles-29413065-8gbhr\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.501816 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/512a131d-22af-40e4-b9ca-a8c04ff844de-secret-volume\") pod \"collect-profiles-29413065-8gbhr\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.512225 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whhw8\" (UniqueName: \"kubernetes.io/projected/512a131d-22af-40e4-b9ca-a8c04ff844de-kube-api-access-whhw8\") pod \"collect-profiles-29413065-8gbhr\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:00 crc kubenswrapper[4787]: I1203 17:45:00.587634 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.025595 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr"] Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.038445 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-xh2lw"] Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.048672 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-acb9-account-create-update-wv7cq"] Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.059959 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-acb9-account-create-update-wv7cq"] Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.069327 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-h5xrc"] Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.080591 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-db-create-xh2lw"] Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.092795 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-h5xrc"] Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.538746 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" event={"ID":"512a131d-22af-40e4-b9ca-a8c04ff844de","Type":"ContainerStarted","Data":"112d87f048f7a3e087df6163dd9fca2bcca1a218762dc78a9eadd669e6d3af1c"} Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.539144 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" event={"ID":"512a131d-22af-40e4-b9ca-a8c04ff844de","Type":"ContainerStarted","Data":"a27362c5eba4f2089a4689d40dc3462aae2a8b77e853f67a2689f20cbf3539dc"} Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.564500 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" podStartSLOduration=1.564479075 podStartE2EDuration="1.564479075s" podCreationTimestamp="2025-12-03 17:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 17:45:01.553678987 +0000 UTC m=+1918.371149966" watchObservedRunningTime="2025-12-03 17:45:01.564479075 +0000 UTC m=+1918.381950034" Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.777793 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2131a38e-a365-46fd-8b7f-b01906edec67" path="/var/lib/kubelet/pods/2131a38e-a365-46fd-8b7f-b01906edec67/volumes" Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.778777 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="40edf4b6-dc92-4d34-bbb2-a3784cf06405" path="/var/lib/kubelet/pods/40edf4b6-dc92-4d34-bbb2-a3784cf06405/volumes" Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.779697 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6c6d0805-aa7a-4038-8245-77e79adbe026" path="/var/lib/kubelet/pods/6c6d0805-aa7a-4038-8245-77e79adbe026/volumes" Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.780647 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ad8fa82a-ce58-4713-900e-f0e333cd8448" path="/var/lib/kubelet/pods/ad8fa82a-ce58-4713-900e-f0e333cd8448/volumes" Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.782410 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3bccb94-fcb5-44fc-819b-e39562a7e57a" path="/var/lib/kubelet/pods/d3bccb94-fcb5-44fc-819b-e39562a7e57a/volumes" Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.783734 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f041eda1-f785-472b-928e-7ab2b901e99e" path="/var/lib/kubelet/pods/f041eda1-f785-472b-928e-7ab2b901e99e/volumes" Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.784788 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb521dcb-0316-4363-983f-1e2800ee2630" path="/var/lib/kubelet/pods/fb521dcb-0316-4363-983f-1e2800ee2630/volumes" Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.845729 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.845776 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:45:01 crc kubenswrapper[4787]: I1203 17:45:01.910238 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:45:02 crc kubenswrapper[4787]: I1203 17:45:02.553682 4787 generic.go:334] "Generic (PLEG): container finished" podID="512a131d-22af-40e4-b9ca-a8c04ff844de" containerID="112d87f048f7a3e087df6163dd9fca2bcca1a218762dc78a9eadd669e6d3af1c" exitCode=0 Dec 03 17:45:02 crc kubenswrapper[4787]: I1203 17:45:02.553781 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" event={"ID":"512a131d-22af-40e4-b9ca-a8c04ff844de","Type":"ContainerDied","Data":"112d87f048f7a3e087df6163dd9fca2bcca1a218762dc78a9eadd669e6d3af1c"} Dec 03 17:45:02 crc kubenswrapper[4787]: I1203 17:45:02.625118 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:45:02 crc kubenswrapper[4787]: I1203 17:45:02.766303 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:45:02 crc kubenswrapper[4787]: E1203 17:45:02.766896 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:45:03 crc kubenswrapper[4787]: I1203 17:45:03.028049 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-fwg8j"] Dec 03 17:45:03 crc kubenswrapper[4787]: I1203 17:45:03.038640 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-fwg8j"] Dec 03 17:45:03 crc kubenswrapper[4787]: I1203 17:45:03.095822 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hf52w"] Dec 03 17:45:03 crc kubenswrapper[4787]: I1203 17:45:03.813460 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73fb0e8b-a595-41f9-b732-7c0984f614e6" path="/var/lib/kubelet/pods/73fb0e8b-a595-41f9-b732-7c0984f614e6/volumes" Dec 03 17:45:03 crc kubenswrapper[4787]: I1203 17:45:03.994284 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.067166 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/512a131d-22af-40e4-b9ca-a8c04ff844de-secret-volume\") pod \"512a131d-22af-40e4-b9ca-a8c04ff844de\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.067274 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/512a131d-22af-40e4-b9ca-a8c04ff844de-config-volume\") pod \"512a131d-22af-40e4-b9ca-a8c04ff844de\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.067320 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whhw8\" (UniqueName: \"kubernetes.io/projected/512a131d-22af-40e4-b9ca-a8c04ff844de-kube-api-access-whhw8\") pod \"512a131d-22af-40e4-b9ca-a8c04ff844de\" (UID: \"512a131d-22af-40e4-b9ca-a8c04ff844de\") " Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.067920 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/512a131d-22af-40e4-b9ca-a8c04ff844de-config-volume" (OuterVolumeSpecName: "config-volume") pod "512a131d-22af-40e4-b9ca-a8c04ff844de" (UID: "512a131d-22af-40e4-b9ca-a8c04ff844de"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.068094 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/512a131d-22af-40e4-b9ca-a8c04ff844de-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.073234 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/512a131d-22af-40e4-b9ca-a8c04ff844de-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "512a131d-22af-40e4-b9ca-a8c04ff844de" (UID: "512a131d-22af-40e4-b9ca-a8c04ff844de"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.073580 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/512a131d-22af-40e4-b9ca-a8c04ff844de-kube-api-access-whhw8" (OuterVolumeSpecName: "kube-api-access-whhw8") pod "512a131d-22af-40e4-b9ca-a8c04ff844de" (UID: "512a131d-22af-40e4-b9ca-a8c04ff844de"). InnerVolumeSpecName "kube-api-access-whhw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.169651 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/512a131d-22af-40e4-b9ca-a8c04ff844de-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.169685 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whhw8\" (UniqueName: \"kubernetes.io/projected/512a131d-22af-40e4-b9ca-a8c04ff844de-kube-api-access-whhw8\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.580927 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hf52w" podUID="e854a964-9d6a-4a35-8048-16dfb61fcefc" containerName="registry-server" containerID="cri-o://2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d" gracePeriod=2 Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.581253 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" event={"ID":"512a131d-22af-40e4-b9ca-a8c04ff844de","Type":"ContainerDied","Data":"a27362c5eba4f2089a4689d40dc3462aae2a8b77e853f67a2689f20cbf3539dc"} Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.581282 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr" Dec 03 17:45:04 crc kubenswrapper[4787]: I1203 17:45:04.581306 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a27362c5eba4f2089a4689d40dc3462aae2a8b77e853f67a2689f20cbf3539dc" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.084430 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.196256 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swwxc\" (UniqueName: \"kubernetes.io/projected/e854a964-9d6a-4a35-8048-16dfb61fcefc-kube-api-access-swwxc\") pod \"e854a964-9d6a-4a35-8048-16dfb61fcefc\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.196533 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-catalog-content\") pod \"e854a964-9d6a-4a35-8048-16dfb61fcefc\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.196577 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-utilities\") pod \"e854a964-9d6a-4a35-8048-16dfb61fcefc\" (UID: \"e854a964-9d6a-4a35-8048-16dfb61fcefc\") " Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.197473 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-utilities" (OuterVolumeSpecName: "utilities") pod "e854a964-9d6a-4a35-8048-16dfb61fcefc" (UID: "e854a964-9d6a-4a35-8048-16dfb61fcefc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.201957 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e854a964-9d6a-4a35-8048-16dfb61fcefc-kube-api-access-swwxc" (OuterVolumeSpecName: "kube-api-access-swwxc") pod "e854a964-9d6a-4a35-8048-16dfb61fcefc" (UID: "e854a964-9d6a-4a35-8048-16dfb61fcefc"). InnerVolumeSpecName "kube-api-access-swwxc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.247301 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e854a964-9d6a-4a35-8048-16dfb61fcefc" (UID: "e854a964-9d6a-4a35-8048-16dfb61fcefc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.299665 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swwxc\" (UniqueName: \"kubernetes.io/projected/e854a964-9d6a-4a35-8048-16dfb61fcefc-kube-api-access-swwxc\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.299916 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.300102 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e854a964-9d6a-4a35-8048-16dfb61fcefc-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.596132 4787 generic.go:334] "Generic (PLEG): container finished" podID="e854a964-9d6a-4a35-8048-16dfb61fcefc" containerID="2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d" exitCode=0 Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.596435 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hf52w" event={"ID":"e854a964-9d6a-4a35-8048-16dfb61fcefc","Type":"ContainerDied","Data":"2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d"} Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.596462 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hf52w" event={"ID":"e854a964-9d6a-4a35-8048-16dfb61fcefc","Type":"ContainerDied","Data":"91672280b49c00973f2f9481a196be0bb451f589735c2e9f496d4e5b8573efbb"} Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.596495 4787 scope.go:117] "RemoveContainer" containerID="2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.596660 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hf52w" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.618500 4787 scope.go:117] "RemoveContainer" containerID="2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.653852 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hf52w"] Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.658137 4787 scope.go:117] "RemoveContainer" containerID="dd23a88e82e2d1f63d43a49dfc968241ba59df7b5608698ab5a83a792aeecd97" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.665961 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hf52w"] Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.696185 4787 scope.go:117] "RemoveContainer" containerID="2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d" Dec 03 17:45:05 crc kubenswrapper[4787]: E1203 17:45:05.696723 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d\": container with ID starting with 2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d not found: ID does not exist" containerID="2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.696876 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d"} err="failed to get container status \"2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d\": rpc error: code = NotFound desc = could not find container \"2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d\": container with ID starting with 2f4b02242d5e03e718e47c9f3f103379501ff7c252630e5a84875c2cf8fcf01d not found: ID does not exist" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.696973 4787 scope.go:117] "RemoveContainer" containerID="2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810" Dec 03 17:45:05 crc kubenswrapper[4787]: E1203 17:45:05.697377 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810\": container with ID starting with 2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810 not found: ID does not exist" containerID="2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.697489 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810"} err="failed to get container status \"2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810\": rpc error: code = NotFound desc = could not find container \"2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810\": container with ID starting with 2a7e48b3f7c67ec10c291dc8810ef6ed6292867a0f588e39d04c7724a6206810 not found: ID does not exist" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.697572 4787 scope.go:117] "RemoveContainer" containerID="dd23a88e82e2d1f63d43a49dfc968241ba59df7b5608698ab5a83a792aeecd97" Dec 03 17:45:05 crc kubenswrapper[4787]: E1203 17:45:05.698050 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd23a88e82e2d1f63d43a49dfc968241ba59df7b5608698ab5a83a792aeecd97\": container with ID starting with dd23a88e82e2d1f63d43a49dfc968241ba59df7b5608698ab5a83a792aeecd97 not found: ID does not exist" containerID="dd23a88e82e2d1f63d43a49dfc968241ba59df7b5608698ab5a83a792aeecd97" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.698077 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd23a88e82e2d1f63d43a49dfc968241ba59df7b5608698ab5a83a792aeecd97"} err="failed to get container status \"dd23a88e82e2d1f63d43a49dfc968241ba59df7b5608698ab5a83a792aeecd97\": rpc error: code = NotFound desc = could not find container \"dd23a88e82e2d1f63d43a49dfc968241ba59df7b5608698ab5a83a792aeecd97\": container with ID starting with dd23a88e82e2d1f63d43a49dfc968241ba59df7b5608698ab5a83a792aeecd97 not found: ID does not exist" Dec 03 17:45:05 crc kubenswrapper[4787]: I1203 17:45:05.784432 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e854a964-9d6a-4a35-8048-16dfb61fcefc" path="/var/lib/kubelet/pods/e854a964-9d6a-4a35-8048-16dfb61fcefc/volumes" Dec 03 17:45:09 crc kubenswrapper[4787]: I1203 17:45:09.037588 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj"] Dec 03 17:45:09 crc kubenswrapper[4787]: I1203 17:45:09.051195 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/mysqld-exporter-960e-account-create-update-8dsds"] Dec 03 17:45:09 crc kubenswrapper[4787]: I1203 17:45:09.061739 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-openstack-cell1-db-create-dhhvj"] Dec 03 17:45:09 crc kubenswrapper[4787]: I1203 17:45:09.071233 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/mysqld-exporter-960e-account-create-update-8dsds"] Dec 03 17:45:09 crc kubenswrapper[4787]: I1203 17:45:09.787279 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ec64d7f-0783-4048-a8ac-baa6de52ed78" path="/var/lib/kubelet/pods/4ec64d7f-0783-4048-a8ac-baa6de52ed78/volumes" Dec 03 17:45:09 crc kubenswrapper[4787]: I1203 17:45:09.788158 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c3b80bcb-cab7-4a0e-a761-ad5942435422" path="/var/lib/kubelet/pods/c3b80bcb-cab7-4a0e-a761-ad5942435422/volumes" Dec 03 17:45:16 crc kubenswrapper[4787]: I1203 17:45:16.766242 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:45:16 crc kubenswrapper[4787]: E1203 17:45:16.767171 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:45:21 crc kubenswrapper[4787]: E1203 17:45:21.132444 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99e1ef0d_6d9c_413a_9d70_e8b0aa3f441b.slice/crio-conmon-7bf32c081a4b702652e0a6f9b5bc7ee04153f83fd34a4535b91f6ade01959fec.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99e1ef0d_6d9c_413a_9d70_e8b0aa3f441b.slice/crio-7bf32c081a4b702652e0a6f9b5bc7ee04153f83fd34a4535b91f6ade01959fec.scope\": RecentStats: unable to find data in memory cache]" Dec 03 17:45:21 crc kubenswrapper[4787]: I1203 17:45:21.777844 4787 generic.go:334] "Generic (PLEG): container finished" podID="99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b" containerID="7bf32c081a4b702652e0a6f9b5bc7ee04153f83fd34a4535b91f6ade01959fec" exitCode=0 Dec 03 17:45:21 crc kubenswrapper[4787]: I1203 17:45:21.781395 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" event={"ID":"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b","Type":"ContainerDied","Data":"7bf32c081a4b702652e0a6f9b5bc7ee04153f83fd34a4535b91f6ade01959fec"} Dec 03 17:45:22 crc kubenswrapper[4787]: I1203 17:45:22.358140 4787 scope.go:117] "RemoveContainer" containerID="9a8da11cb9d3a4532da1b97a323c979a02e6f2fbd5e26b282773c81d9a63d9af" Dec 03 17:45:22 crc kubenswrapper[4787]: I1203 17:45:22.446452 4787 scope.go:117] "RemoveContainer" containerID="58d22e695149a0fb7682e4f032435ed3494a52daba772e2c13d278ccb01a0466" Dec 03 17:45:22 crc kubenswrapper[4787]: I1203 17:45:22.488224 4787 scope.go:117] "RemoveContainer" containerID="31c94fc96a99da5e5eb901fcd8b315d2db0356c0e5377be64b601229f62c5736" Dec 03 17:45:22 crc kubenswrapper[4787]: I1203 17:45:22.537250 4787 scope.go:117] "RemoveContainer" containerID="84641faf23ffc7fef426487234f9466e3374f556bd45a42a5246c36715f69d97" Dec 03 17:45:22 crc kubenswrapper[4787]: I1203 17:45:22.581475 4787 scope.go:117] "RemoveContainer" containerID="b0d33df450a3d278beed7b67c25c2b638db42b0e4b8191ae8d5f4335dfc63e14" Dec 03 17:45:22 crc kubenswrapper[4787]: I1203 17:45:22.627194 4787 scope.go:117] "RemoveContainer" containerID="d92ea270829886b26a447aca6fe702d5b2d50b1c12d2e00cc73679bb7ca315ae" Dec 03 17:45:22 crc kubenswrapper[4787]: I1203 17:45:22.683876 4787 scope.go:117] "RemoveContainer" containerID="4e177c0d754e8bb4b61123598ea157a8cf8c780e1d967e48235b4d539e011a3e" Dec 03 17:45:22 crc kubenswrapper[4787]: I1203 17:45:22.709216 4787 scope.go:117] "RemoveContainer" containerID="cf2d60b2fed518e2538ab724aeb9860633b53fb0f42497d14dfeb0c62a7955c0" Dec 03 17:45:22 crc kubenswrapper[4787]: I1203 17:45:22.733251 4787 scope.go:117] "RemoveContainer" containerID="021fa014eff70f032132ca3763ca662441515e52086cce85320c771165ecb316" Dec 03 17:45:22 crc kubenswrapper[4787]: I1203 17:45:22.763404 4787 scope.go:117] "RemoveContainer" containerID="a6379aca724ec8746c8ea782a4bbfb03129c8da4126ccbedb7ff77b67336416e" Dec 03 17:45:22 crc kubenswrapper[4787]: I1203 17:45:22.797721 4787 scope.go:117] "RemoveContainer" containerID="93248ed1ec7fb5c91bd3039d139a9b291650631c029183d0fca52f09a2c128fa" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.173192 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.283561 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-inventory\") pod \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.283658 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgjjq\" (UniqueName: \"kubernetes.io/projected/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-kube-api-access-bgjjq\") pod \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.283800 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-bootstrap-combined-ca-bundle\") pod \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.283997 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-ssh-key\") pod \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\" (UID: \"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b\") " Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.289399 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-kube-api-access-bgjjq" (OuterVolumeSpecName: "kube-api-access-bgjjq") pod "99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b" (UID: "99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b"). InnerVolumeSpecName "kube-api-access-bgjjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.289823 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b" (UID: "99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.313346 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b" (UID: "99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.318154 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-inventory" (OuterVolumeSpecName: "inventory") pod "99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b" (UID: "99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.386076 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.386106 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.386117 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgjjq\" (UniqueName: \"kubernetes.io/projected/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-kube-api-access-bgjjq\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.386129 4787 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.832281 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" event={"ID":"99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b","Type":"ContainerDied","Data":"56781e71c20a1bb81255d0a686d9c83dff6998cba5b2e5aaf2e82e8ea7f85c6e"} Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.832358 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56781e71c20a1bb81255d0a686d9c83dff6998cba5b2e5aaf2e82e8ea7f85c6e" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.832388 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.895817 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69"] Dec 03 17:45:23 crc kubenswrapper[4787]: E1203 17:45:23.896282 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.896301 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 17:45:23 crc kubenswrapper[4787]: E1203 17:45:23.896316 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e854a964-9d6a-4a35-8048-16dfb61fcefc" containerName="extract-content" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.896323 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e854a964-9d6a-4a35-8048-16dfb61fcefc" containerName="extract-content" Dec 03 17:45:23 crc kubenswrapper[4787]: E1203 17:45:23.896339 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e854a964-9d6a-4a35-8048-16dfb61fcefc" containerName="extract-utilities" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.896345 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e854a964-9d6a-4a35-8048-16dfb61fcefc" containerName="extract-utilities" Dec 03 17:45:23 crc kubenswrapper[4787]: E1203 17:45:23.896353 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="512a131d-22af-40e4-b9ca-a8c04ff844de" containerName="collect-profiles" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.896358 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="512a131d-22af-40e4-b9ca-a8c04ff844de" containerName="collect-profiles" Dec 03 17:45:23 crc kubenswrapper[4787]: E1203 17:45:23.896402 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e854a964-9d6a-4a35-8048-16dfb61fcefc" containerName="registry-server" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.896409 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e854a964-9d6a-4a35-8048-16dfb61fcefc" containerName="registry-server" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.896591 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.896614 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e854a964-9d6a-4a35-8048-16dfb61fcefc" containerName="registry-server" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.896625 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="512a131d-22af-40e4-b9ca-a8c04ff844de" containerName="collect-profiles" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.897364 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.900729 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.900931 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.900863 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.901341 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:45:23 crc kubenswrapper[4787]: I1203 17:45:23.907994 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69"] Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.001451 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rtc69\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.001640 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l2qw\" (UniqueName: \"kubernetes.io/projected/2aabe7bb-06d8-44b6-a293-d228d461c1ab-kube-api-access-9l2qw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rtc69\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.001681 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rtc69\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.103442 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rtc69\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.103522 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rtc69\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.103715 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l2qw\" (UniqueName: \"kubernetes.io/projected/2aabe7bb-06d8-44b6-a293-d228d461c1ab-kube-api-access-9l2qw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rtc69\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.107745 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rtc69\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.107923 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rtc69\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.121333 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l2qw\" (UniqueName: \"kubernetes.io/projected/2aabe7bb-06d8-44b6-a293-d228d461c1ab-kube-api-access-9l2qw\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-rtc69\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.225336 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.761673 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69"] Dec 03 17:45:24 crc kubenswrapper[4787]: I1203 17:45:24.842819 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" event={"ID":"2aabe7bb-06d8-44b6-a293-d228d461c1ab","Type":"ContainerStarted","Data":"df7013487d176cd3e7815c9c34ed775def52b8efcf4eea576b30e49f34c60861"} Dec 03 17:45:25 crc kubenswrapper[4787]: I1203 17:45:25.857594 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" event={"ID":"2aabe7bb-06d8-44b6-a293-d228d461c1ab","Type":"ContainerStarted","Data":"ad1416f9bce630e669d0ff1bc345a80417673ceb40c0f5973c46e096978e87b1"} Dec 03 17:45:25 crc kubenswrapper[4787]: I1203 17:45:25.875938 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" podStartSLOduration=2.464106312 podStartE2EDuration="2.875902619s" podCreationTimestamp="2025-12-03 17:45:23 +0000 UTC" firstStartedPulling="2025-12-03 17:45:24.766611703 +0000 UTC m=+1941.584082662" lastFinishedPulling="2025-12-03 17:45:25.17840796 +0000 UTC m=+1941.995878969" observedRunningTime="2025-12-03 17:45:25.874162952 +0000 UTC m=+1942.691633931" watchObservedRunningTime="2025-12-03 17:45:25.875902619 +0000 UTC m=+1942.693373578" Dec 03 17:45:27 crc kubenswrapper[4787]: I1203 17:45:27.765721 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:45:27 crc kubenswrapper[4787]: E1203 17:45:27.766623 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:45:31 crc kubenswrapper[4787]: I1203 17:45:31.039315 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-wcnd6"] Dec 03 17:45:31 crc kubenswrapper[4787]: I1203 17:45:31.049540 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-wcnd6"] Dec 03 17:45:31 crc kubenswrapper[4787]: I1203 17:45:31.779440 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="948f9158-3e6d-4a5d-a8d4-1c9480a85f90" path="/var/lib/kubelet/pods/948f9158-3e6d-4a5d-a8d4-1c9480a85f90/volumes" Dec 03 17:45:41 crc kubenswrapper[4787]: I1203 17:45:41.765941 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:45:41 crc kubenswrapper[4787]: E1203 17:45:41.766820 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.073420 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-f622-account-create-update-mwk99"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.087294 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-f4q94"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.100409 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-zxhg9"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.114681 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-create-9r7gn"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.126333 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-d920-account-create-update-nqrk5"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.139840 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-81ba-account-create-update-qfjj9"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.161384 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-8e2a-account-create-update-9z9kk"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.173227 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-4477n"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.184103 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-f622-account-create-update-mwk99"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.193198 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-8e2a-account-create-update-9z9kk"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.201989 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-zxhg9"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.212119 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-f4q94"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.221174 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-d920-account-create-update-nqrk5"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.230371 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-4477n"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.239401 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-81ba-account-create-update-qfjj9"] Dec 03 17:45:44 crc kubenswrapper[4787]: I1203 17:45:44.248815 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-create-9r7gn"] Dec 03 17:45:45 crc kubenswrapper[4787]: I1203 17:45:45.780996 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3f36e27a-5a90-4f53-8b13-e42e134471f8" path="/var/lib/kubelet/pods/3f36e27a-5a90-4f53-8b13-e42e134471f8/volumes" Dec 03 17:45:45 crc kubenswrapper[4787]: I1203 17:45:45.781824 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4ec2c795-db12-41f1-a2cc-536471d01938" path="/var/lib/kubelet/pods/4ec2c795-db12-41f1-a2cc-536471d01938/volumes" Dec 03 17:45:45 crc kubenswrapper[4787]: I1203 17:45:45.782590 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="766cf6f2-1add-4b7e-a5d6-16facb6782dd" path="/var/lib/kubelet/pods/766cf6f2-1add-4b7e-a5d6-16facb6782dd/volumes" Dec 03 17:45:45 crc kubenswrapper[4787]: I1203 17:45:45.783390 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ee1f7fe-90fd-4289-b280-f48d0076ce09" path="/var/lib/kubelet/pods/8ee1f7fe-90fd-4289-b280-f48d0076ce09/volumes" Dec 03 17:45:45 crc kubenswrapper[4787]: I1203 17:45:45.785370 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b643325f-a8b6-4f76-ad35-52e0c4fd525e" path="/var/lib/kubelet/pods/b643325f-a8b6-4f76-ad35-52e0c4fd525e/volumes" Dec 03 17:45:45 crc kubenswrapper[4787]: I1203 17:45:45.786121 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c0405ea9-d37c-40a5-98aa-86ed0d81c288" path="/var/lib/kubelet/pods/c0405ea9-d37c-40a5-98aa-86ed0d81c288/volumes" Dec 03 17:45:45 crc kubenswrapper[4787]: I1203 17:45:45.786825 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5" path="/var/lib/kubelet/pods/d1864b1c-9a12-4cdd-9f5f-9d0d29e29ef5/volumes" Dec 03 17:45:45 crc kubenswrapper[4787]: I1203 17:45:45.788209 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f43e786a-e9ff-4ca7-a0d6-f9b9ce928679" path="/var/lib/kubelet/pods/f43e786a-e9ff-4ca7-a0d6-f9b9ce928679/volumes" Dec 03 17:45:49 crc kubenswrapper[4787]: I1203 17:45:49.040236 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-sx9j7"] Dec 03 17:45:49 crc kubenswrapper[4787]: I1203 17:45:49.057915 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-sx9j7"] Dec 03 17:45:49 crc kubenswrapper[4787]: I1203 17:45:49.787770 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dadfdc4d-30a6-4edd-88a4-8835c6e433c0" path="/var/lib/kubelet/pods/dadfdc4d-30a6-4edd-88a4-8835c6e433c0/volumes" Dec 03 17:45:55 crc kubenswrapper[4787]: I1203 17:45:55.766963 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:45:55 crc kubenswrapper[4787]: E1203 17:45:55.767994 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:46:07 crc kubenswrapper[4787]: I1203 17:46:07.766481 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:46:07 crc kubenswrapper[4787]: E1203 17:46:07.767597 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:46:20 crc kubenswrapper[4787]: I1203 17:46:20.766308 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:46:21 crc kubenswrapper[4787]: I1203 17:46:21.465195 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"c2571bed769c5e81a57d85c403092e3b76ad0831c285bb44f058f97347d8398e"} Dec 03 17:46:23 crc kubenswrapper[4787]: I1203 17:46:23.058986 4787 scope.go:117] "RemoveContainer" containerID="5d26d36c88fe30d95249b94ec41893577fc809ea40eef42fce8ab575360bf0dc" Dec 03 17:46:23 crc kubenswrapper[4787]: I1203 17:46:23.096437 4787 scope.go:117] "RemoveContainer" containerID="bbfbd7081f97a5b2201c799c4903d708f81a0e9cfc311ecd112967c9c896d002" Dec 03 17:46:23 crc kubenswrapper[4787]: I1203 17:46:23.160555 4787 scope.go:117] "RemoveContainer" containerID="47331be765582c7502e1bb5e32f953f0bbab28d27d64505f9058a7a31f75862d" Dec 03 17:46:23 crc kubenswrapper[4787]: I1203 17:46:23.203316 4787 scope.go:117] "RemoveContainer" containerID="77f034cd17d81b7795902c5d95791ed7f11491750435688747e326e98228c78b" Dec 03 17:46:23 crc kubenswrapper[4787]: I1203 17:46:23.286708 4787 scope.go:117] "RemoveContainer" containerID="9f237298f174bca5da3b59c792602109df5fa64d6589447b51dba505bdf71de7" Dec 03 17:46:23 crc kubenswrapper[4787]: I1203 17:46:23.325208 4787 scope.go:117] "RemoveContainer" containerID="6f2e1cb7a6767cf507693445b32f7919569248618d8d0ea7829b4ac3c85d02c9" Dec 03 17:46:23 crc kubenswrapper[4787]: I1203 17:46:23.362276 4787 scope.go:117] "RemoveContainer" containerID="ac282b9d9d2a8599bfff0d29214bcdaa84ddda0651f2a9d8b262a9922a498891" Dec 03 17:46:23 crc kubenswrapper[4787]: I1203 17:46:23.383893 4787 scope.go:117] "RemoveContainer" containerID="42f5d5e3986f90c800097a400e9327f1fdd2e107d2cc8fab37c993e8910fbf8d" Dec 03 17:46:23 crc kubenswrapper[4787]: I1203 17:46:23.402485 4787 scope.go:117] "RemoveContainer" containerID="a8b56fb0ac3b3b1adea50394e76383617e9788f6ca5f9a25b53d165e1a15a749" Dec 03 17:46:23 crc kubenswrapper[4787]: I1203 17:46:23.423622 4787 scope.go:117] "RemoveContainer" containerID="a1f0db31f1451a1e3dc6428e9a1678b70a9ab04ce4dea0d76978ed185d13a824" Dec 03 17:46:24 crc kubenswrapper[4787]: I1203 17:46:24.064293 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-zrgzg"] Dec 03 17:46:24 crc kubenswrapper[4787]: I1203 17:46:24.087482 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-zrgzg"] Dec 03 17:46:24 crc kubenswrapper[4787]: I1203 17:46:24.098166 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-2r5xr"] Dec 03 17:46:24 crc kubenswrapper[4787]: I1203 17:46:24.106646 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-2r5xr"] Dec 03 17:46:25 crc kubenswrapper[4787]: I1203 17:46:25.040940 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-qqt8b"] Dec 03 17:46:25 crc kubenswrapper[4787]: I1203 17:46:25.060558 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-qqt8b"] Dec 03 17:46:25 crc kubenswrapper[4787]: I1203 17:46:25.788535 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0308bd88-1943-4aea-a041-b5e7bd9e64f3" path="/var/lib/kubelet/pods/0308bd88-1943-4aea-a041-b5e7bd9e64f3/volumes" Dec 03 17:46:25 crc kubenswrapper[4787]: I1203 17:46:25.789935 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a39f222-78f1-4f59-a557-bcfee7f3104a" path="/var/lib/kubelet/pods/7a39f222-78f1-4f59-a557-bcfee7f3104a/volumes" Dec 03 17:46:25 crc kubenswrapper[4787]: I1203 17:46:25.792053 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af99de84-794c-4fc3-9582-6007e90c71f3" path="/var/lib/kubelet/pods/af99de84-794c-4fc3-9582-6007e90c71f3/volumes" Dec 03 17:46:35 crc kubenswrapper[4787]: I1203 17:46:35.660206 4787 generic.go:334] "Generic (PLEG): container finished" podID="2aabe7bb-06d8-44b6-a293-d228d461c1ab" containerID="ad1416f9bce630e669d0ff1bc345a80417673ceb40c0f5973c46e096978e87b1" exitCode=0 Dec 03 17:46:35 crc kubenswrapper[4787]: I1203 17:46:35.660281 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" event={"ID":"2aabe7bb-06d8-44b6-a293-d228d461c1ab","Type":"ContainerDied","Data":"ad1416f9bce630e669d0ff1bc345a80417673ceb40c0f5973c46e096978e87b1"} Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.145344 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.225840 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-ssh-key\") pod \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.226233 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-inventory\") pod \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.226398 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l2qw\" (UniqueName: \"kubernetes.io/projected/2aabe7bb-06d8-44b6-a293-d228d461c1ab-kube-api-access-9l2qw\") pod \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\" (UID: \"2aabe7bb-06d8-44b6-a293-d228d461c1ab\") " Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.232484 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2aabe7bb-06d8-44b6-a293-d228d461c1ab-kube-api-access-9l2qw" (OuterVolumeSpecName: "kube-api-access-9l2qw") pod "2aabe7bb-06d8-44b6-a293-d228d461c1ab" (UID: "2aabe7bb-06d8-44b6-a293-d228d461c1ab"). InnerVolumeSpecName "kube-api-access-9l2qw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.275826 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-inventory" (OuterVolumeSpecName: "inventory") pod "2aabe7bb-06d8-44b6-a293-d228d461c1ab" (UID: "2aabe7bb-06d8-44b6-a293-d228d461c1ab"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.275850 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "2aabe7bb-06d8-44b6-a293-d228d461c1ab" (UID: "2aabe7bb-06d8-44b6-a293-d228d461c1ab"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.329614 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.329649 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/2aabe7bb-06d8-44b6-a293-d228d461c1ab-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.329662 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l2qw\" (UniqueName: \"kubernetes.io/projected/2aabe7bb-06d8-44b6-a293-d228d461c1ab-kube-api-access-9l2qw\") on node \"crc\" DevicePath \"\"" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.702291 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" event={"ID":"2aabe7bb-06d8-44b6-a293-d228d461c1ab","Type":"ContainerDied","Data":"df7013487d176cd3e7815c9c34ed775def52b8efcf4eea576b30e49f34c60861"} Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.702355 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df7013487d176cd3e7815c9c34ed775def52b8efcf4eea576b30e49f34c60861" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.702445 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.787855 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk"] Dec 03 17:46:37 crc kubenswrapper[4787]: E1203 17:46:37.788587 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2aabe7bb-06d8-44b6-a293-d228d461c1ab" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.788605 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2aabe7bb-06d8-44b6-a293-d228d461c1ab" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.788883 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2aabe7bb-06d8-44b6-a293-d228d461c1ab" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.789744 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.792745 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.794076 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.794697 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.795072 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.804467 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk"] Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.942556 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.942597 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6rlh\" (UniqueName: \"kubernetes.io/projected/e0400992-8418-4339-9162-f788984931e1-kube-api-access-p6rlh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:37 crc kubenswrapper[4787]: I1203 17:46:37.942689 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:38 crc kubenswrapper[4787]: I1203 17:46:38.044939 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:38 crc kubenswrapper[4787]: I1203 17:46:38.045130 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6rlh\" (UniqueName: \"kubernetes.io/projected/e0400992-8418-4339-9162-f788984931e1-kube-api-access-p6rlh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:38 crc kubenswrapper[4787]: I1203 17:46:38.045151 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:38 crc kubenswrapper[4787]: I1203 17:46:38.051249 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:38 crc kubenswrapper[4787]: I1203 17:46:38.051379 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:38 crc kubenswrapper[4787]: I1203 17:46:38.062003 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6rlh\" (UniqueName: \"kubernetes.io/projected/e0400992-8418-4339-9162-f788984931e1-kube-api-access-p6rlh\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:38 crc kubenswrapper[4787]: I1203 17:46:38.111242 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:38 crc kubenswrapper[4787]: I1203 17:46:38.678889 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk"] Dec 03 17:46:38 crc kubenswrapper[4787]: I1203 17:46:38.690474 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:46:38 crc kubenswrapper[4787]: I1203 17:46:38.713528 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" event={"ID":"e0400992-8418-4339-9162-f788984931e1","Type":"ContainerStarted","Data":"091d152b2be9696e4f67d027e5ac67c62235a4b6d084e08673568fec12789c66"} Dec 03 17:46:39 crc kubenswrapper[4787]: I1203 17:46:39.723562 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" event={"ID":"e0400992-8418-4339-9162-f788984931e1","Type":"ContainerStarted","Data":"506f9832b7028efce81cadd12518b375f1a755602fdc73cfddeb622111b0cbc4"} Dec 03 17:46:39 crc kubenswrapper[4787]: I1203 17:46:39.747615 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" podStartSLOduration=2.291812953 podStartE2EDuration="2.747591025s" podCreationTimestamp="2025-12-03 17:46:37 +0000 UTC" firstStartedPulling="2025-12-03 17:46:38.690154856 +0000 UTC m=+2015.507625815" lastFinishedPulling="2025-12-03 17:46:39.145932918 +0000 UTC m=+2015.963403887" observedRunningTime="2025-12-03 17:46:39.745620642 +0000 UTC m=+2016.563091601" watchObservedRunningTime="2025-12-03 17:46:39.747591025 +0000 UTC m=+2016.565062014" Dec 03 17:46:40 crc kubenswrapper[4787]: I1203 17:46:40.047818 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-fkrgl"] Dec 03 17:46:40 crc kubenswrapper[4787]: I1203 17:46:40.060158 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-fkrgl"] Dec 03 17:46:41 crc kubenswrapper[4787]: I1203 17:46:41.777381 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="372bef16-5eab-41db-acbf-07274bf3f985" path="/var/lib/kubelet/pods/372bef16-5eab-41db-acbf-07274bf3f985/volumes" Dec 03 17:46:44 crc kubenswrapper[4787]: I1203 17:46:44.803686 4787 generic.go:334] "Generic (PLEG): container finished" podID="e0400992-8418-4339-9162-f788984931e1" containerID="506f9832b7028efce81cadd12518b375f1a755602fdc73cfddeb622111b0cbc4" exitCode=0 Dec 03 17:46:44 crc kubenswrapper[4787]: I1203 17:46:44.803761 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" event={"ID":"e0400992-8418-4339-9162-f788984931e1","Type":"ContainerDied","Data":"506f9832b7028efce81cadd12518b375f1a755602fdc73cfddeb622111b0cbc4"} Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.270553 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.328210 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p6rlh\" (UniqueName: \"kubernetes.io/projected/e0400992-8418-4339-9162-f788984931e1-kube-api-access-p6rlh\") pod \"e0400992-8418-4339-9162-f788984931e1\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.328305 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-ssh-key\") pod \"e0400992-8418-4339-9162-f788984931e1\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.328575 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-inventory\") pod \"e0400992-8418-4339-9162-f788984931e1\" (UID: \"e0400992-8418-4339-9162-f788984931e1\") " Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.334887 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0400992-8418-4339-9162-f788984931e1-kube-api-access-p6rlh" (OuterVolumeSpecName: "kube-api-access-p6rlh") pod "e0400992-8418-4339-9162-f788984931e1" (UID: "e0400992-8418-4339-9162-f788984931e1"). InnerVolumeSpecName "kube-api-access-p6rlh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.359966 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-inventory" (OuterVolumeSpecName: "inventory") pod "e0400992-8418-4339-9162-f788984931e1" (UID: "e0400992-8418-4339-9162-f788984931e1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.360415 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "e0400992-8418-4339-9162-f788984931e1" (UID: "e0400992-8418-4339-9162-f788984931e1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.432762 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.432890 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p6rlh\" (UniqueName: \"kubernetes.io/projected/e0400992-8418-4339-9162-f788984931e1-kube-api-access-p6rlh\") on node \"crc\" DevicePath \"\"" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.432955 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/e0400992-8418-4339-9162-f788984931e1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.826252 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" event={"ID":"e0400992-8418-4339-9162-f788984931e1","Type":"ContainerDied","Data":"091d152b2be9696e4f67d027e5ac67c62235a4b6d084e08673568fec12789c66"} Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.826496 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="091d152b2be9696e4f67d027e5ac67c62235a4b6d084e08673568fec12789c66" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.826313 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.958773 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p"] Dec 03 17:46:46 crc kubenswrapper[4787]: E1203 17:46:46.959268 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0400992-8418-4339-9162-f788984931e1" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.959290 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0400992-8418-4339-9162-f788984931e1" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.959547 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0400992-8418-4339-9162-f788984931e1" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.960373 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.963151 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.963461 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.963636 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.965397 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:46:46 crc kubenswrapper[4787]: I1203 17:46:46.974569 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p"] Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.045884 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9ws7\" (UniqueName: \"kubernetes.io/projected/6aaee49b-7923-4882-9db0-28ad5d07dea9-kube-api-access-t9ws7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g975p\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.045983 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g975p\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.046067 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g975p\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.148045 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9ws7\" (UniqueName: \"kubernetes.io/projected/6aaee49b-7923-4882-9db0-28ad5d07dea9-kube-api-access-t9ws7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g975p\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.148132 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g975p\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.148190 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g975p\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.153531 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g975p\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.153949 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g975p\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.165611 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9ws7\" (UniqueName: \"kubernetes.io/projected/6aaee49b-7923-4882-9db0-28ad5d07dea9-kube-api-access-t9ws7\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-g975p\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.278936 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.818992 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p"] Dec 03 17:46:47 crc kubenswrapper[4787]: I1203 17:46:47.835720 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" event={"ID":"6aaee49b-7923-4882-9db0-28ad5d07dea9","Type":"ContainerStarted","Data":"0c7d40810d4b310504f135acb7bee1540d4bdb4fb4e5d165df83422f4046f65b"} Dec 03 17:46:48 crc kubenswrapper[4787]: I1203 17:46:48.851367 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" event={"ID":"6aaee49b-7923-4882-9db0-28ad5d07dea9","Type":"ContainerStarted","Data":"3b79134240d56cbb27fe03ad1a79b5e97a9eb8fc8fe6f48d84c0f14dcd94d033"} Dec 03 17:46:48 crc kubenswrapper[4787]: I1203 17:46:48.883479 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" podStartSLOduration=2.471547504 podStartE2EDuration="2.883456039s" podCreationTimestamp="2025-12-03 17:46:46 +0000 UTC" firstStartedPulling="2025-12-03 17:46:47.824868538 +0000 UTC m=+2024.642339497" lastFinishedPulling="2025-12-03 17:46:48.236777073 +0000 UTC m=+2025.054248032" observedRunningTime="2025-12-03 17:46:48.870819819 +0000 UTC m=+2025.688290818" watchObservedRunningTime="2025-12-03 17:46:48.883456039 +0000 UTC m=+2025.700927008" Dec 03 17:46:50 crc kubenswrapper[4787]: I1203 17:46:50.043466 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-zmx58"] Dec 03 17:46:50 crc kubenswrapper[4787]: I1203 17:46:50.057794 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-zmx58"] Dec 03 17:46:51 crc kubenswrapper[4787]: I1203 17:46:51.776773 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90" path="/var/lib/kubelet/pods/cfab6daa-4ee6-4b9a-b60c-4a553cbe4d90/volumes" Dec 03 17:47:23 crc kubenswrapper[4787]: I1203 17:47:23.639057 4787 scope.go:117] "RemoveContainer" containerID="5ea8872fcab9bad9d38cfa5fe6bf0b6a99b91830dbdf9c5b24932474ef68a99f" Dec 03 17:47:23 crc kubenswrapper[4787]: I1203 17:47:23.687908 4787 scope.go:117] "RemoveContainer" containerID="4e391d1dfb528dbb106ba0a803f077358aa28a2f3ca76df74e46062fc61ffb2b" Dec 03 17:47:23 crc kubenswrapper[4787]: I1203 17:47:23.751116 4787 scope.go:117] "RemoveContainer" containerID="b9dfc7706b50afd1dd55f3a25a7644cf617bc170fecd695147bb42b76c618dc5" Dec 03 17:47:23 crc kubenswrapper[4787]: I1203 17:47:23.826410 4787 scope.go:117] "RemoveContainer" containerID="59b595c210a822d7a538658e8fea2885d20ffeacd789a83f7a59ca43f7527951" Dec 03 17:47:23 crc kubenswrapper[4787]: I1203 17:47:23.873550 4787 scope.go:117] "RemoveContainer" containerID="0551230c6ebe89aad9bca04ef713444f4765ff4aae8f2b2e78bd5faef6d51db3" Dec 03 17:47:25 crc kubenswrapper[4787]: I1203 17:47:25.055339 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-a7a8-account-create-update-72nlv"] Dec 03 17:47:25 crc kubenswrapper[4787]: I1203 17:47:25.068414 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-a7a8-account-create-update-72nlv"] Dec 03 17:47:25 crc kubenswrapper[4787]: I1203 17:47:25.793373 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="39227401-0c97-4fd7-8510-615e22ca73d9" path="/var/lib/kubelet/pods/39227401-0c97-4fd7-8510-615e22ca73d9/volumes" Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.036967 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-3101-account-create-update-g8j4t"] Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.047959 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-hj85k"] Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.060477 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-3101-account-create-update-g8j4t"] Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.070526 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-hj85k"] Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.078423 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-91eb-account-create-update-59l2z"] Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.086493 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-91eb-account-create-update-59l2z"] Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.094411 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-lcm5g"] Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.101819 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-lcm5g"] Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.785348 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="244e40a3-b432-4971-bdf5-f070ab9891b0" path="/var/lib/kubelet/pods/244e40a3-b432-4971-bdf5-f070ab9891b0/volumes" Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.786598 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2ea7d1a3-de18-4283-bd8e-2be80f82eb9a" path="/var/lib/kubelet/pods/2ea7d1a3-de18-4283-bd8e-2be80f82eb9a/volumes" Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.787556 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6371d53b-b035-4a6e-82b5-c5774ffc8de1" path="/var/lib/kubelet/pods/6371d53b-b035-4a6e-82b5-c5774ffc8de1/volumes" Dec 03 17:47:27 crc kubenswrapper[4787]: I1203 17:47:27.788494 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="75bff842-ba6b-4f84-bff6-054b7292d82e" path="/var/lib/kubelet/pods/75bff842-ba6b-4f84-bff6-054b7292d82e/volumes" Dec 03 17:47:28 crc kubenswrapper[4787]: I1203 17:47:28.034469 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-6dhs9"] Dec 03 17:47:28 crc kubenswrapper[4787]: I1203 17:47:28.045116 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-6dhs9"] Dec 03 17:47:28 crc kubenswrapper[4787]: I1203 17:47:28.301113 4787 generic.go:334] "Generic (PLEG): container finished" podID="6aaee49b-7923-4882-9db0-28ad5d07dea9" containerID="3b79134240d56cbb27fe03ad1a79b5e97a9eb8fc8fe6f48d84c0f14dcd94d033" exitCode=0 Dec 03 17:47:28 crc kubenswrapper[4787]: I1203 17:47:28.301480 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" event={"ID":"6aaee49b-7923-4882-9db0-28ad5d07dea9","Type":"ContainerDied","Data":"3b79134240d56cbb27fe03ad1a79b5e97a9eb8fc8fe6f48d84c0f14dcd94d033"} Dec 03 17:47:29 crc kubenswrapper[4787]: I1203 17:47:29.780297 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1a23593c-7762-4639-8b2f-2cdd49b0c728" path="/var/lib/kubelet/pods/1a23593c-7762-4639-8b2f-2cdd49b0c728/volumes" Dec 03 17:47:29 crc kubenswrapper[4787]: I1203 17:47:29.813867 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:47:29 crc kubenswrapper[4787]: I1203 17:47:29.932384 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-inventory\") pod \"6aaee49b-7923-4882-9db0-28ad5d07dea9\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " Dec 03 17:47:29 crc kubenswrapper[4787]: I1203 17:47:29.932711 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9ws7\" (UniqueName: \"kubernetes.io/projected/6aaee49b-7923-4882-9db0-28ad5d07dea9-kube-api-access-t9ws7\") pod \"6aaee49b-7923-4882-9db0-28ad5d07dea9\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " Dec 03 17:47:29 crc kubenswrapper[4787]: I1203 17:47:29.933123 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-ssh-key\") pod \"6aaee49b-7923-4882-9db0-28ad5d07dea9\" (UID: \"6aaee49b-7923-4882-9db0-28ad5d07dea9\") " Dec 03 17:47:29 crc kubenswrapper[4787]: I1203 17:47:29.936905 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6aaee49b-7923-4882-9db0-28ad5d07dea9-kube-api-access-t9ws7" (OuterVolumeSpecName: "kube-api-access-t9ws7") pod "6aaee49b-7923-4882-9db0-28ad5d07dea9" (UID: "6aaee49b-7923-4882-9db0-28ad5d07dea9"). InnerVolumeSpecName "kube-api-access-t9ws7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:47:29 crc kubenswrapper[4787]: I1203 17:47:29.958081 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-inventory" (OuterVolumeSpecName: "inventory") pod "6aaee49b-7923-4882-9db0-28ad5d07dea9" (UID: "6aaee49b-7923-4882-9db0-28ad5d07dea9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:47:29 crc kubenswrapper[4787]: I1203 17:47:29.969799 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6aaee49b-7923-4882-9db0-28ad5d07dea9" (UID: "6aaee49b-7923-4882-9db0-28ad5d07dea9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.036853 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.036884 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9ws7\" (UniqueName: \"kubernetes.io/projected/6aaee49b-7923-4882-9db0-28ad5d07dea9-kube-api-access-t9ws7\") on node \"crc\" DevicePath \"\"" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.036896 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6aaee49b-7923-4882-9db0-28ad5d07dea9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.323669 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" event={"ID":"6aaee49b-7923-4882-9db0-28ad5d07dea9","Type":"ContainerDied","Data":"0c7d40810d4b310504f135acb7bee1540d4bdb4fb4e5d165df83422f4046f65b"} Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.323706 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c7d40810d4b310504f135acb7bee1540d4bdb4fb4e5d165df83422f4046f65b" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.323836 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.423777 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d"] Dec 03 17:47:30 crc kubenswrapper[4787]: E1203 17:47:30.424216 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6aaee49b-7923-4882-9db0-28ad5d07dea9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.424240 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6aaee49b-7923-4882-9db0-28ad5d07dea9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.424477 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6aaee49b-7923-4882-9db0-28ad5d07dea9" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.425281 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.428551 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.428757 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.428835 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.429171 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.454370 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d"] Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.548791 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vd8vc\" (UniqueName: \"kubernetes.io/projected/bb00d591-6e35-41bd-be71-a8a8190946b1-kube-api-access-vd8vc\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.549203 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.549672 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.651620 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vd8vc\" (UniqueName: \"kubernetes.io/projected/bb00d591-6e35-41bd-be71-a8a8190946b1-kube-api-access-vd8vc\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.652381 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.652796 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.656883 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.658902 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.670405 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vd8vc\" (UniqueName: \"kubernetes.io/projected/bb00d591-6e35-41bd-be71-a8a8190946b1-kube-api-access-vd8vc\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:30 crc kubenswrapper[4787]: I1203 17:47:30.746528 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:31 crc kubenswrapper[4787]: I1203 17:47:31.296889 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d"] Dec 03 17:47:31 crc kubenswrapper[4787]: I1203 17:47:31.336004 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" event={"ID":"bb00d591-6e35-41bd-be71-a8a8190946b1","Type":"ContainerStarted","Data":"cc3b0548b04eaa58920a8dc6c5b28d99675b524e0693da7cd68fe7cdb119e91a"} Dec 03 17:47:32 crc kubenswrapper[4787]: I1203 17:47:32.348922 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" event={"ID":"bb00d591-6e35-41bd-be71-a8a8190946b1","Type":"ContainerStarted","Data":"ed0e7cff92f75edd35950004a414ef9da7a31f4b1eaa2bccd6b52cd682582c65"} Dec 03 17:47:32 crc kubenswrapper[4787]: I1203 17:47:32.375948 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" podStartSLOduration=1.80990151 podStartE2EDuration="2.37592675s" podCreationTimestamp="2025-12-03 17:47:30 +0000 UTC" firstStartedPulling="2025-12-03 17:47:31.300896729 +0000 UTC m=+2068.118367728" lastFinishedPulling="2025-12-03 17:47:31.866921999 +0000 UTC m=+2068.684392968" observedRunningTime="2025-12-03 17:47:32.36697727 +0000 UTC m=+2069.184448249" watchObservedRunningTime="2025-12-03 17:47:32.37592675 +0000 UTC m=+2069.193397719" Dec 03 17:47:36 crc kubenswrapper[4787]: I1203 17:47:36.401187 4787 generic.go:334] "Generic (PLEG): container finished" podID="bb00d591-6e35-41bd-be71-a8a8190946b1" containerID="ed0e7cff92f75edd35950004a414ef9da7a31f4b1eaa2bccd6b52cd682582c65" exitCode=0 Dec 03 17:47:36 crc kubenswrapper[4787]: I1203 17:47:36.401863 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" event={"ID":"bb00d591-6e35-41bd-be71-a8a8190946b1","Type":"ContainerDied","Data":"ed0e7cff92f75edd35950004a414ef9da7a31f4b1eaa2bccd6b52cd682582c65"} Dec 03 17:47:37 crc kubenswrapper[4787]: I1203 17:47:37.963521 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.024718 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vd8vc\" (UniqueName: \"kubernetes.io/projected/bb00d591-6e35-41bd-be71-a8a8190946b1-kube-api-access-vd8vc\") pod \"bb00d591-6e35-41bd-be71-a8a8190946b1\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.024799 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-ssh-key\") pod \"bb00d591-6e35-41bd-be71-a8a8190946b1\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.024869 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-inventory\") pod \"bb00d591-6e35-41bd-be71-a8a8190946b1\" (UID: \"bb00d591-6e35-41bd-be71-a8a8190946b1\") " Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.031447 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb00d591-6e35-41bd-be71-a8a8190946b1-kube-api-access-vd8vc" (OuterVolumeSpecName: "kube-api-access-vd8vc") pod "bb00d591-6e35-41bd-be71-a8a8190946b1" (UID: "bb00d591-6e35-41bd-be71-a8a8190946b1"). InnerVolumeSpecName "kube-api-access-vd8vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.069127 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "bb00d591-6e35-41bd-be71-a8a8190946b1" (UID: "bb00d591-6e35-41bd-be71-a8a8190946b1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.076524 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-inventory" (OuterVolumeSpecName: "inventory") pod "bb00d591-6e35-41bd-be71-a8a8190946b1" (UID: "bb00d591-6e35-41bd-be71-a8a8190946b1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.127165 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vd8vc\" (UniqueName: \"kubernetes.io/projected/bb00d591-6e35-41bd-be71-a8a8190946b1-kube-api-access-vd8vc\") on node \"crc\" DevicePath \"\"" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.127391 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.127453 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/bb00d591-6e35-41bd-be71-a8a8190946b1-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.431970 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" event={"ID":"bb00d591-6e35-41bd-be71-a8a8190946b1","Type":"ContainerDied","Data":"cc3b0548b04eaa58920a8dc6c5b28d99675b524e0693da7cd68fe7cdb119e91a"} Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.432366 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc3b0548b04eaa58920a8dc6c5b28d99675b524e0693da7cd68fe7cdb119e91a" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.432115 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.498964 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt"] Dec 03 17:47:38 crc kubenswrapper[4787]: E1203 17:47:38.499478 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb00d591-6e35-41bd-be71-a8a8190946b1" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.499502 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb00d591-6e35-41bd-be71-a8a8190946b1" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.499743 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb00d591-6e35-41bd-be71-a8a8190946b1" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.500611 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.507330 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.507460 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.507868 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.508627 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.525711 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt"] Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.637579 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jjskt\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.637692 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gsdv\" (UniqueName: \"kubernetes.io/projected/66d0ee2f-607b-4a9c-9808-4200fef6c906-kube-api-access-2gsdv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jjskt\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.637733 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jjskt\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.740166 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gsdv\" (UniqueName: \"kubernetes.io/projected/66d0ee2f-607b-4a9c-9808-4200fef6c906-kube-api-access-2gsdv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jjskt\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.740223 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jjskt\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.740400 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jjskt\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.760905 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jjskt\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.760932 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jjskt\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.765697 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gsdv\" (UniqueName: \"kubernetes.io/projected/66d0ee2f-607b-4a9c-9808-4200fef6c906-kube-api-access-2gsdv\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-jjskt\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:47:38 crc kubenswrapper[4787]: I1203 17:47:38.820360 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:47:39 crc kubenswrapper[4787]: I1203 17:47:39.389607 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt"] Dec 03 17:47:39 crc kubenswrapper[4787]: I1203 17:47:39.444702 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" event={"ID":"66d0ee2f-607b-4a9c-9808-4200fef6c906","Type":"ContainerStarted","Data":"096f89fe97b34dfac99b1b59d703277490461088093cdfc5cd4859d871dd35d2"} Dec 03 17:47:40 crc kubenswrapper[4787]: I1203 17:47:40.475338 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" event={"ID":"66d0ee2f-607b-4a9c-9808-4200fef6c906","Type":"ContainerStarted","Data":"9b616255df00e40b54d812bd65dc8d85465f9f84eb0b597a78083c6c41152ece"} Dec 03 17:47:40 crc kubenswrapper[4787]: I1203 17:47:40.503072 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" podStartSLOduration=2.066074136 podStartE2EDuration="2.503057083s" podCreationTimestamp="2025-12-03 17:47:38 +0000 UTC" firstStartedPulling="2025-12-03 17:47:39.394166303 +0000 UTC m=+2076.211637282" lastFinishedPulling="2025-12-03 17:47:39.83114926 +0000 UTC m=+2076.648620229" observedRunningTime="2025-12-03 17:47:40.502882398 +0000 UTC m=+2077.320353367" watchObservedRunningTime="2025-12-03 17:47:40.503057083 +0000 UTC m=+2077.320528042" Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.740285 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-krhh4"] Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.743105 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.756894 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krhh4"] Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.800598 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-catalog-content\") pod \"redhat-operators-krhh4\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.800725 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4cns\" (UniqueName: \"kubernetes.io/projected/8b27786a-3d37-4df1-99b6-a6e9317d0971-kube-api-access-d4cns\") pod \"redhat-operators-krhh4\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.800851 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-utilities\") pod \"redhat-operators-krhh4\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.903432 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-catalog-content\") pod \"redhat-operators-krhh4\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.903805 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4cns\" (UniqueName: \"kubernetes.io/projected/8b27786a-3d37-4df1-99b6-a6e9317d0971-kube-api-access-d4cns\") pod \"redhat-operators-krhh4\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.903941 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-utilities\") pod \"redhat-operators-krhh4\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.904312 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-catalog-content\") pod \"redhat-operators-krhh4\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.904312 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-utilities\") pod \"redhat-operators-krhh4\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:47:59 crc kubenswrapper[4787]: I1203 17:47:59.926221 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4cns\" (UniqueName: \"kubernetes.io/projected/8b27786a-3d37-4df1-99b6-a6e9317d0971-kube-api-access-d4cns\") pod \"redhat-operators-krhh4\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:48:00 crc kubenswrapper[4787]: I1203 17:48:00.066685 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:48:00 crc kubenswrapper[4787]: I1203 17:48:00.535130 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-krhh4"] Dec 03 17:48:00 crc kubenswrapper[4787]: W1203 17:48:00.537809 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8b27786a_3d37_4df1_99b6_a6e9317d0971.slice/crio-49c1c4e1e375606f38e2b46eb5d77bce116e750968edb46d70f6e2aee9a0d3b0 WatchSource:0}: Error finding container 49c1c4e1e375606f38e2b46eb5d77bce116e750968edb46d70f6e2aee9a0d3b0: Status 404 returned error can't find the container with id 49c1c4e1e375606f38e2b46eb5d77bce116e750968edb46d70f6e2aee9a0d3b0 Dec 03 17:48:00 crc kubenswrapper[4787]: I1203 17:48:00.698604 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhh4" event={"ID":"8b27786a-3d37-4df1-99b6-a6e9317d0971","Type":"ContainerStarted","Data":"49c1c4e1e375606f38e2b46eb5d77bce116e750968edb46d70f6e2aee9a0d3b0"} Dec 03 17:48:01 crc kubenswrapper[4787]: I1203 17:48:01.045087 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l2pkv"] Dec 03 17:48:01 crc kubenswrapper[4787]: I1203 17:48:01.056739 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-l2pkv"] Dec 03 17:48:01 crc kubenswrapper[4787]: I1203 17:48:01.721326 4787 generic.go:334] "Generic (PLEG): container finished" podID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerID="84b8fd38c005ef68ff1ced04feeb376614008b241e74cbdb9cfddb1967b0d685" exitCode=0 Dec 03 17:48:01 crc kubenswrapper[4787]: I1203 17:48:01.721600 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhh4" event={"ID":"8b27786a-3d37-4df1-99b6-a6e9317d0971","Type":"ContainerDied","Data":"84b8fd38c005ef68ff1ced04feeb376614008b241e74cbdb9cfddb1967b0d685"} Dec 03 17:48:01 crc kubenswrapper[4787]: I1203 17:48:01.805389 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="131d89aa-86f5-4ddb-a388-f2a938319ce2" path="/var/lib/kubelet/pods/131d89aa-86f5-4ddb-a388-f2a938319ce2/volumes" Dec 03 17:48:02 crc kubenswrapper[4787]: I1203 17:48:02.732921 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhh4" event={"ID":"8b27786a-3d37-4df1-99b6-a6e9317d0971","Type":"ContainerStarted","Data":"7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61"} Dec 03 17:48:06 crc kubenswrapper[4787]: I1203 17:48:06.771175 4787 generic.go:334] "Generic (PLEG): container finished" podID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerID="7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61" exitCode=0 Dec 03 17:48:06 crc kubenswrapper[4787]: I1203 17:48:06.771216 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhh4" event={"ID":"8b27786a-3d37-4df1-99b6-a6e9317d0971","Type":"ContainerDied","Data":"7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61"} Dec 03 17:48:07 crc kubenswrapper[4787]: I1203 17:48:07.784464 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhh4" event={"ID":"8b27786a-3d37-4df1-99b6-a6e9317d0971","Type":"ContainerStarted","Data":"bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29"} Dec 03 17:48:07 crc kubenswrapper[4787]: I1203 17:48:07.804869 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-krhh4" podStartSLOduration=3.281027069 podStartE2EDuration="8.804846455s" podCreationTimestamp="2025-12-03 17:47:59 +0000 UTC" firstStartedPulling="2025-12-03 17:48:01.724638407 +0000 UTC m=+2098.542109386" lastFinishedPulling="2025-12-03 17:48:07.248457773 +0000 UTC m=+2104.065928772" observedRunningTime="2025-12-03 17:48:07.802410759 +0000 UTC m=+2104.619881718" watchObservedRunningTime="2025-12-03 17:48:07.804846455 +0000 UTC m=+2104.622317414" Dec 03 17:48:10 crc kubenswrapper[4787]: I1203 17:48:10.066881 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:48:10 crc kubenswrapper[4787]: I1203 17:48:10.067366 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:48:11 crc kubenswrapper[4787]: I1203 17:48:11.117842 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-krhh4" podUID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerName="registry-server" probeResult="failure" output=< Dec 03 17:48:11 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 17:48:11 crc kubenswrapper[4787]: > Dec 03 17:48:20 crc kubenswrapper[4787]: I1203 17:48:20.067041 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-0c2d-account-create-update-5kwbn"] Dec 03 17:48:20 crc kubenswrapper[4787]: I1203 17:48:20.087130 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-create-rb7vv"] Dec 03 17:48:20 crc kubenswrapper[4787]: I1203 17:48:20.099183 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-0c2d-account-create-update-5kwbn"] Dec 03 17:48:20 crc kubenswrapper[4787]: I1203 17:48:20.109414 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-create-rb7vv"] Dec 03 17:48:20 crc kubenswrapper[4787]: I1203 17:48:20.121775 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:48:20 crc kubenswrapper[4787]: I1203 17:48:20.171595 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:48:20 crc kubenswrapper[4787]: I1203 17:48:20.359350 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krhh4"] Dec 03 17:48:21 crc kubenswrapper[4787]: I1203 17:48:21.778748 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a476357-691f-402d-b70e-f81a8f37b27e" path="/var/lib/kubelet/pods/6a476357-691f-402d-b70e-f81a8f37b27e/volumes" Dec 03 17:48:21 crc kubenswrapper[4787]: I1203 17:48:21.779671 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4a7a56f-da6c-41e4-8147-ab175d3a0285" path="/var/lib/kubelet/pods/b4a7a56f-da6c-41e4-8147-ab175d3a0285/volumes" Dec 03 17:48:21 crc kubenswrapper[4787]: I1203 17:48:21.935732 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-krhh4" podUID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerName="registry-server" containerID="cri-o://bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29" gracePeriod=2 Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.443620 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.514581 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-catalog-content\") pod \"8b27786a-3d37-4df1-99b6-a6e9317d0971\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.514683 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-utilities\") pod \"8b27786a-3d37-4df1-99b6-a6e9317d0971\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.514714 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4cns\" (UniqueName: \"kubernetes.io/projected/8b27786a-3d37-4df1-99b6-a6e9317d0971-kube-api-access-d4cns\") pod \"8b27786a-3d37-4df1-99b6-a6e9317d0971\" (UID: \"8b27786a-3d37-4df1-99b6-a6e9317d0971\") " Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.516056 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-utilities" (OuterVolumeSpecName: "utilities") pod "8b27786a-3d37-4df1-99b6-a6e9317d0971" (UID: "8b27786a-3d37-4df1-99b6-a6e9317d0971"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.521659 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b27786a-3d37-4df1-99b6-a6e9317d0971-kube-api-access-d4cns" (OuterVolumeSpecName: "kube-api-access-d4cns") pod "8b27786a-3d37-4df1-99b6-a6e9317d0971" (UID: "8b27786a-3d37-4df1-99b6-a6e9317d0971"). InnerVolumeSpecName "kube-api-access-d4cns". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.617618 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.617662 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4cns\" (UniqueName: \"kubernetes.io/projected/8b27786a-3d37-4df1-99b6-a6e9317d0971-kube-api-access-d4cns\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.649050 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b27786a-3d37-4df1-99b6-a6e9317d0971" (UID: "8b27786a-3d37-4df1-99b6-a6e9317d0971"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.720078 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b27786a-3d37-4df1-99b6-a6e9317d0971-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.948661 4787 generic.go:334] "Generic (PLEG): container finished" podID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerID="bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29" exitCode=0 Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.948699 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhh4" event={"ID":"8b27786a-3d37-4df1-99b6-a6e9317d0971","Type":"ContainerDied","Data":"bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29"} Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.948736 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-krhh4" Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.948757 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-krhh4" event={"ID":"8b27786a-3d37-4df1-99b6-a6e9317d0971","Type":"ContainerDied","Data":"49c1c4e1e375606f38e2b46eb5d77bce116e750968edb46d70f6e2aee9a0d3b0"} Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.948785 4787 scope.go:117] "RemoveContainer" containerID="bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29" Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.984476 4787 scope.go:117] "RemoveContainer" containerID="7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61" Dec 03 17:48:22 crc kubenswrapper[4787]: I1203 17:48:22.996328 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-krhh4"] Dec 03 17:48:23 crc kubenswrapper[4787]: I1203 17:48:23.011354 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-krhh4"] Dec 03 17:48:23 crc kubenswrapper[4787]: I1203 17:48:23.031177 4787 scope.go:117] "RemoveContainer" containerID="84b8fd38c005ef68ff1ced04feeb376614008b241e74cbdb9cfddb1967b0d685" Dec 03 17:48:23 crc kubenswrapper[4787]: I1203 17:48:23.075948 4787 scope.go:117] "RemoveContainer" containerID="bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29" Dec 03 17:48:23 crc kubenswrapper[4787]: E1203 17:48:23.076734 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29\": container with ID starting with bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29 not found: ID does not exist" containerID="bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29" Dec 03 17:48:23 crc kubenswrapper[4787]: I1203 17:48:23.076762 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29"} err="failed to get container status \"bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29\": rpc error: code = NotFound desc = could not find container \"bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29\": container with ID starting with bb8027f445b3e52fcca909bb81fda3c0d17b3db401492adbbeca5869e6d79b29 not found: ID does not exist" Dec 03 17:48:23 crc kubenswrapper[4787]: I1203 17:48:23.076906 4787 scope.go:117] "RemoveContainer" containerID="7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61" Dec 03 17:48:23 crc kubenswrapper[4787]: E1203 17:48:23.077599 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61\": container with ID starting with 7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61 not found: ID does not exist" containerID="7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61" Dec 03 17:48:23 crc kubenswrapper[4787]: I1203 17:48:23.077628 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61"} err="failed to get container status \"7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61\": rpc error: code = NotFound desc = could not find container \"7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61\": container with ID starting with 7450d8829fcf39161e4c36b3e69c73df86549f0732af4f84050e061a7d4d7f61 not found: ID does not exist" Dec 03 17:48:23 crc kubenswrapper[4787]: I1203 17:48:23.077644 4787 scope.go:117] "RemoveContainer" containerID="84b8fd38c005ef68ff1ced04feeb376614008b241e74cbdb9cfddb1967b0d685" Dec 03 17:48:23 crc kubenswrapper[4787]: E1203 17:48:23.078297 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84b8fd38c005ef68ff1ced04feeb376614008b241e74cbdb9cfddb1967b0d685\": container with ID starting with 84b8fd38c005ef68ff1ced04feeb376614008b241e74cbdb9cfddb1967b0d685 not found: ID does not exist" containerID="84b8fd38c005ef68ff1ced04feeb376614008b241e74cbdb9cfddb1967b0d685" Dec 03 17:48:23 crc kubenswrapper[4787]: I1203 17:48:23.078316 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84b8fd38c005ef68ff1ced04feeb376614008b241e74cbdb9cfddb1967b0d685"} err="failed to get container status \"84b8fd38c005ef68ff1ced04feeb376614008b241e74cbdb9cfddb1967b0d685\": rpc error: code = NotFound desc = could not find container \"84b8fd38c005ef68ff1ced04feeb376614008b241e74cbdb9cfddb1967b0d685\": container with ID starting with 84b8fd38c005ef68ff1ced04feeb376614008b241e74cbdb9cfddb1967b0d685 not found: ID does not exist" Dec 03 17:48:23 crc kubenswrapper[4787]: I1203 17:48:23.777153 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b27786a-3d37-4df1-99b6-a6e9317d0971" path="/var/lib/kubelet/pods/8b27786a-3d37-4df1-99b6-a6e9317d0971/volumes" Dec 03 17:48:24 crc kubenswrapper[4787]: I1203 17:48:24.026243 4787 scope.go:117] "RemoveContainer" containerID="094e601df92d88190a1f624d9ac4d8030d1e73bd06959edc6e27b20fe473f3e2" Dec 03 17:48:24 crc kubenswrapper[4787]: I1203 17:48:24.051105 4787 scope.go:117] "RemoveContainer" containerID="4e810cda512274f9a0e0575876cb1477e0a06af4d0c0f1d365cda6ba6262f8fb" Dec 03 17:48:24 crc kubenswrapper[4787]: I1203 17:48:24.127812 4787 scope.go:117] "RemoveContainer" containerID="4db0b4029701f28ef38c1a9528668be4f59569f70282bda4cc0240c8f55fcb83" Dec 03 17:48:24 crc kubenswrapper[4787]: I1203 17:48:24.155354 4787 scope.go:117] "RemoveContainer" containerID="32cf70c8718d0cf030ccd84af5935a7264a1072b0d92369eb54dfd538ddedf04" Dec 03 17:48:24 crc kubenswrapper[4787]: I1203 17:48:24.198789 4787 scope.go:117] "RemoveContainer" containerID="5f8bd0c2316a0b41c49138685d20b95d4b5a8b7d5ba8721f05f2b058295c00e6" Dec 03 17:48:24 crc kubenswrapper[4787]: I1203 17:48:24.251307 4787 scope.go:117] "RemoveContainer" containerID="f88de925b2be6062cba8c51bb2e7a43a3762824e7bcff3dc75c995498933e662" Dec 03 17:48:24 crc kubenswrapper[4787]: I1203 17:48:24.304310 4787 scope.go:117] "RemoveContainer" containerID="0f939d4c6f7e73d503b443ea7afeeb178113b0790db90fe2bbf9eb309a6948da" Dec 03 17:48:24 crc kubenswrapper[4787]: I1203 17:48:24.347430 4787 scope.go:117] "RemoveContainer" containerID="20cf456e5428a9bed8dec9ce1a32971183064f80563c71f03e150f7a0cb67760" Dec 03 17:48:24 crc kubenswrapper[4787]: I1203 17:48:24.369596 4787 scope.go:117] "RemoveContainer" containerID="14e0483968ab6ca747903bdaf85df573297632c5106490083f10f40a4fbd5d88" Dec 03 17:48:32 crc kubenswrapper[4787]: I1203 17:48:32.046359 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-zq27q"] Dec 03 17:48:32 crc kubenswrapper[4787]: I1203 17:48:32.074432 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-zq27q"] Dec 03 17:48:32 crc kubenswrapper[4787]: I1203 17:48:32.078717 4787 generic.go:334] "Generic (PLEG): container finished" podID="66d0ee2f-607b-4a9c-9808-4200fef6c906" containerID="9b616255df00e40b54d812bd65dc8d85465f9f84eb0b597a78083c6c41152ece" exitCode=0 Dec 03 17:48:32 crc kubenswrapper[4787]: I1203 17:48:32.078761 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" event={"ID":"66d0ee2f-607b-4a9c-9808-4200fef6c906","Type":"ContainerDied","Data":"9b616255df00e40b54d812bd65dc8d85465f9f84eb0b597a78083c6c41152ece"} Dec 03 17:48:33 crc kubenswrapper[4787]: I1203 17:48:33.543568 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:48:33 crc kubenswrapper[4787]: I1203 17:48:33.664277 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-inventory\") pod \"66d0ee2f-607b-4a9c-9808-4200fef6c906\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " Dec 03 17:48:33 crc kubenswrapper[4787]: I1203 17:48:33.664550 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2gsdv\" (UniqueName: \"kubernetes.io/projected/66d0ee2f-607b-4a9c-9808-4200fef6c906-kube-api-access-2gsdv\") pod \"66d0ee2f-607b-4a9c-9808-4200fef6c906\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " Dec 03 17:48:33 crc kubenswrapper[4787]: I1203 17:48:33.664582 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-ssh-key\") pod \"66d0ee2f-607b-4a9c-9808-4200fef6c906\" (UID: \"66d0ee2f-607b-4a9c-9808-4200fef6c906\") " Dec 03 17:48:33 crc kubenswrapper[4787]: I1203 17:48:33.670202 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/66d0ee2f-607b-4a9c-9808-4200fef6c906-kube-api-access-2gsdv" (OuterVolumeSpecName: "kube-api-access-2gsdv") pod "66d0ee2f-607b-4a9c-9808-4200fef6c906" (UID: "66d0ee2f-607b-4a9c-9808-4200fef6c906"). InnerVolumeSpecName "kube-api-access-2gsdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:48:33 crc kubenswrapper[4787]: I1203 17:48:33.700305 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "66d0ee2f-607b-4a9c-9808-4200fef6c906" (UID: "66d0ee2f-607b-4a9c-9808-4200fef6c906"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:48:33 crc kubenswrapper[4787]: I1203 17:48:33.700917 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-inventory" (OuterVolumeSpecName: "inventory") pod "66d0ee2f-607b-4a9c-9808-4200fef6c906" (UID: "66d0ee2f-607b-4a9c-9808-4200fef6c906"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:48:33 crc kubenswrapper[4787]: I1203 17:48:33.772255 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:33 crc kubenswrapper[4787]: I1203 17:48:33.772649 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2gsdv\" (UniqueName: \"kubernetes.io/projected/66d0ee2f-607b-4a9c-9808-4200fef6c906-kube-api-access-2gsdv\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:33 crc kubenswrapper[4787]: I1203 17:48:33.772660 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/66d0ee2f-607b-4a9c-9808-4200fef6c906-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:33 crc kubenswrapper[4787]: I1203 17:48:33.780758 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5f00a92-19b0-4f7b-a3b8-66b3e720fab3" path="/var/lib/kubelet/pods/a5f00a92-19b0-4f7b-a3b8-66b3e720fab3/volumes" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.103312 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" event={"ID":"66d0ee2f-607b-4a9c-9808-4200fef6c906","Type":"ContainerDied","Data":"096f89fe97b34dfac99b1b59d703277490461088093cdfc5cd4859d871dd35d2"} Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.103355 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="096f89fe97b34dfac99b1b59d703277490461088093cdfc5cd4859d871dd35d2" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.103412 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.198130 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-z9phj"] Dec 03 17:48:34 crc kubenswrapper[4787]: E1203 17:48:34.198665 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerName="extract-utilities" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.198691 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerName="extract-utilities" Dec 03 17:48:34 crc kubenswrapper[4787]: E1203 17:48:34.198734 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="66d0ee2f-607b-4a9c-9808-4200fef6c906" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.198744 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="66d0ee2f-607b-4a9c-9808-4200fef6c906" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:48:34 crc kubenswrapper[4787]: E1203 17:48:34.198759 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerName="registry-server" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.198766 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerName="registry-server" Dec 03 17:48:34 crc kubenswrapper[4787]: E1203 17:48:34.198786 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerName="extract-content" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.198794 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerName="extract-content" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.199045 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="66d0ee2f-607b-4a9c-9808-4200fef6c906" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.199071 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b27786a-3d37-4df1-99b6-a6e9317d0971" containerName="registry-server" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.200011 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.202602 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.202804 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.202609 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.203093 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.214008 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-z9phj"] Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.283425 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-z9phj\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.283661 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwdn7\" (UniqueName: \"kubernetes.io/projected/e3e4d4bd-145d-417d-a792-b74eed16e8e0-kube-api-access-mwdn7\") pod \"ssh-known-hosts-edpm-deployment-z9phj\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.283733 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-z9phj\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.385804 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-z9phj\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.386061 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-z9phj\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.386210 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwdn7\" (UniqueName: \"kubernetes.io/projected/e3e4d4bd-145d-417d-a792-b74eed16e8e0-kube-api-access-mwdn7\") pod \"ssh-known-hosts-edpm-deployment-z9phj\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.391205 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-z9phj\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.393524 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-z9phj\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.417296 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwdn7\" (UniqueName: \"kubernetes.io/projected/e3e4d4bd-145d-417d-a792-b74eed16e8e0-kube-api-access-mwdn7\") pod \"ssh-known-hosts-edpm-deployment-z9phj\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:34 crc kubenswrapper[4787]: I1203 17:48:34.535144 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:35 crc kubenswrapper[4787]: I1203 17:48:35.121516 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-z9phj"] Dec 03 17:48:36 crc kubenswrapper[4787]: I1203 17:48:36.128333 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" event={"ID":"e3e4d4bd-145d-417d-a792-b74eed16e8e0","Type":"ContainerStarted","Data":"8dec503b458e8e8f0422ac38bb3eec52b51c3d02d46b7dafd0d516e3ef462128"} Dec 03 17:48:36 crc kubenswrapper[4787]: I1203 17:48:36.129962 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" event={"ID":"e3e4d4bd-145d-417d-a792-b74eed16e8e0","Type":"ContainerStarted","Data":"950e1b01a09dfad99f9ffd51619e743e7367a7860e99f339e22b657dff599dfa"} Dec 03 17:48:37 crc kubenswrapper[4787]: I1203 17:48:37.043422 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" podStartSLOduration=2.618723648 podStartE2EDuration="3.043398866s" podCreationTimestamp="2025-12-03 17:48:34 +0000 UTC" firstStartedPulling="2025-12-03 17:48:35.129062879 +0000 UTC m=+2131.946533838" lastFinishedPulling="2025-12-03 17:48:35.553738087 +0000 UTC m=+2132.371209056" observedRunningTime="2025-12-03 17:48:36.156605546 +0000 UTC m=+2132.974076515" watchObservedRunningTime="2025-12-03 17:48:37.043398866 +0000 UTC m=+2133.860869825" Dec 03 17:48:37 crc kubenswrapper[4787]: I1203 17:48:37.048639 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-45bq2"] Dec 03 17:48:37 crc kubenswrapper[4787]: I1203 17:48:37.059807 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-45bq2"] Dec 03 17:48:37 crc kubenswrapper[4787]: I1203 17:48:37.778154 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e43d4111-655b-4843-bdc1-1465d8c5a050" path="/var/lib/kubelet/pods/e43d4111-655b-4843-bdc1-1465d8c5a050/volumes" Dec 03 17:48:43 crc kubenswrapper[4787]: I1203 17:48:43.200892 4787 generic.go:334] "Generic (PLEG): container finished" podID="e3e4d4bd-145d-417d-a792-b74eed16e8e0" containerID="8dec503b458e8e8f0422ac38bb3eec52b51c3d02d46b7dafd0d516e3ef462128" exitCode=0 Dec 03 17:48:43 crc kubenswrapper[4787]: I1203 17:48:43.200961 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" event={"ID":"e3e4d4bd-145d-417d-a792-b74eed16e8e0","Type":"ContainerDied","Data":"8dec503b458e8e8f0422ac38bb3eec52b51c3d02d46b7dafd0d516e3ef462128"} Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.568887 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-29l7v"] Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.572045 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.583985 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-29l7v"] Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.612129 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpmvs\" (UniqueName: \"kubernetes.io/projected/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-kube-api-access-bpmvs\") pod \"redhat-marketplace-29l7v\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.612282 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-utilities\") pod \"redhat-marketplace-29l7v\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.612387 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-catalog-content\") pod \"redhat-marketplace-29l7v\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.715035 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-utilities\") pod \"redhat-marketplace-29l7v\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.715464 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-catalog-content\") pod \"redhat-marketplace-29l7v\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.715546 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-utilities\") pod \"redhat-marketplace-29l7v\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.715788 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpmvs\" (UniqueName: \"kubernetes.io/projected/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-kube-api-access-bpmvs\") pod \"redhat-marketplace-29l7v\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.715931 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-catalog-content\") pod \"redhat-marketplace-29l7v\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.737915 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.740707 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpmvs\" (UniqueName: \"kubernetes.io/projected/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-kube-api-access-bpmvs\") pod \"redhat-marketplace-29l7v\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.817306 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-inventory-0\") pod \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.817929 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwdn7\" (UniqueName: \"kubernetes.io/projected/e3e4d4bd-145d-417d-a792-b74eed16e8e0-kube-api-access-mwdn7\") pod \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.818246 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-ssh-key-openstack-edpm-ipam\") pod \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\" (UID: \"e3e4d4bd-145d-417d-a792-b74eed16e8e0\") " Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.821670 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e3e4d4bd-145d-417d-a792-b74eed16e8e0-kube-api-access-mwdn7" (OuterVolumeSpecName: "kube-api-access-mwdn7") pod "e3e4d4bd-145d-417d-a792-b74eed16e8e0" (UID: "e3e4d4bd-145d-417d-a792-b74eed16e8e0"). InnerVolumeSpecName "kube-api-access-mwdn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.845735 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "e3e4d4bd-145d-417d-a792-b74eed16e8e0" (UID: "e3e4d4bd-145d-417d-a792-b74eed16e8e0"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.848105 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "e3e4d4bd-145d-417d-a792-b74eed16e8e0" (UID: "e3e4d4bd-145d-417d-a792-b74eed16e8e0"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.901888 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.921179 4787 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.921215 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwdn7\" (UniqueName: \"kubernetes.io/projected/e3e4d4bd-145d-417d-a792-b74eed16e8e0-kube-api-access-mwdn7\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:44 crc kubenswrapper[4787]: I1203 17:48:44.921228 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/e3e4d4bd-145d-417d-a792-b74eed16e8e0-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.221480 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" event={"ID":"e3e4d4bd-145d-417d-a792-b74eed16e8e0","Type":"ContainerDied","Data":"950e1b01a09dfad99f9ffd51619e743e7367a7860e99f339e22b657dff599dfa"} Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.221775 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="950e1b01a09dfad99f9ffd51619e743e7367a7860e99f339e22b657dff599dfa" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.221553 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-z9phj" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.283788 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2"] Dec 03 17:48:45 crc kubenswrapper[4787]: E1203 17:48:45.285865 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e3e4d4bd-145d-417d-a792-b74eed16e8e0" containerName="ssh-known-hosts-edpm-deployment" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.285966 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e3e4d4bd-145d-417d-a792-b74eed16e8e0" containerName="ssh-known-hosts-edpm-deployment" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.286438 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e3e4d4bd-145d-417d-a792-b74eed16e8e0" containerName="ssh-known-hosts-edpm-deployment" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.287535 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.290278 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.291092 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.291271 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.291529 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.295820 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2"] Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.337537 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r5sh2\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.337607 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r5sh2\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.337658 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jst8n\" (UniqueName: \"kubernetes.io/projected/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-kube-api-access-jst8n\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r5sh2\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.360313 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-29l7v"] Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.439623 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jst8n\" (UniqueName: \"kubernetes.io/projected/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-kube-api-access-jst8n\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r5sh2\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.439798 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r5sh2\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.439834 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r5sh2\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.447407 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r5sh2\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.447580 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r5sh2\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.456589 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jst8n\" (UniqueName: \"kubernetes.io/projected/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-kube-api-access-jst8n\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-r5sh2\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:45 crc kubenswrapper[4787]: I1203 17:48:45.605470 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:46 crc kubenswrapper[4787]: I1203 17:48:46.122658 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2"] Dec 03 17:48:46 crc kubenswrapper[4787]: I1203 17:48:46.232010 4787 generic.go:334] "Generic (PLEG): container finished" podID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" containerID="1f7f4253b84502eefb3103e56a8189f82f34c96a8a612eca85441deec62152ed" exitCode=0 Dec 03 17:48:46 crc kubenswrapper[4787]: I1203 17:48:46.232091 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29l7v" event={"ID":"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb","Type":"ContainerDied","Data":"1f7f4253b84502eefb3103e56a8189f82f34c96a8a612eca85441deec62152ed"} Dec 03 17:48:46 crc kubenswrapper[4787]: I1203 17:48:46.232138 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29l7v" event={"ID":"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb","Type":"ContainerStarted","Data":"41739f6c8cdff8579bed569af1864b8a956761b352674003b34940457e8731d9"} Dec 03 17:48:46 crc kubenswrapper[4787]: I1203 17:48:46.234317 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" event={"ID":"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08","Type":"ContainerStarted","Data":"029224a694b97fdeda51ba6f1b5927e0ca46db1ae5c8f483115eb7de16ef08c1"} Dec 03 17:48:47 crc kubenswrapper[4787]: I1203 17:48:47.248077 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29l7v" event={"ID":"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb","Type":"ContainerStarted","Data":"bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223"} Dec 03 17:48:47 crc kubenswrapper[4787]: I1203 17:48:47.249524 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" event={"ID":"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08","Type":"ContainerStarted","Data":"1667e55d178e8f38a692c5ea94bde08a6200d42064af14191e9ac90e82e74946"} Dec 03 17:48:47 crc kubenswrapper[4787]: I1203 17:48:47.291508 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" podStartSLOduration=1.8413143170000001 podStartE2EDuration="2.291490349s" podCreationTimestamp="2025-12-03 17:48:45 +0000 UTC" firstStartedPulling="2025-12-03 17:48:46.122597338 +0000 UTC m=+2142.940068297" lastFinishedPulling="2025-12-03 17:48:46.57277337 +0000 UTC m=+2143.390244329" observedRunningTime="2025-12-03 17:48:47.281291335 +0000 UTC m=+2144.098762294" watchObservedRunningTime="2025-12-03 17:48:47.291490349 +0000 UTC m=+2144.108961298" Dec 03 17:48:48 crc kubenswrapper[4787]: I1203 17:48:48.262616 4787 generic.go:334] "Generic (PLEG): container finished" podID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" containerID="bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223" exitCode=0 Dec 03 17:48:48 crc kubenswrapper[4787]: I1203 17:48:48.262704 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29l7v" event={"ID":"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb","Type":"ContainerDied","Data":"bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223"} Dec 03 17:48:48 crc kubenswrapper[4787]: I1203 17:48:48.989928 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:48:48 crc kubenswrapper[4787]: I1203 17:48:48.990221 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:48:49 crc kubenswrapper[4787]: I1203 17:48:49.275149 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29l7v" event={"ID":"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb","Type":"ContainerStarted","Data":"25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a"} Dec 03 17:48:49 crc kubenswrapper[4787]: I1203 17:48:49.307118 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-29l7v" podStartSLOduration=2.796124913 podStartE2EDuration="5.307097963s" podCreationTimestamp="2025-12-03 17:48:44 +0000 UTC" firstStartedPulling="2025-12-03 17:48:46.233673919 +0000 UTC m=+2143.051144878" lastFinishedPulling="2025-12-03 17:48:48.744646959 +0000 UTC m=+2145.562117928" observedRunningTime="2025-12-03 17:48:49.298533433 +0000 UTC m=+2146.116004392" watchObservedRunningTime="2025-12-03 17:48:49.307097963 +0000 UTC m=+2146.124568922" Dec 03 17:48:54 crc kubenswrapper[4787]: I1203 17:48:54.902848 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:54 crc kubenswrapper[4787]: I1203 17:48:54.903553 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:54 crc kubenswrapper[4787]: I1203 17:48:54.968232 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:55 crc kubenswrapper[4787]: I1203 17:48:55.335254 4787 generic.go:334] "Generic (PLEG): container finished" podID="dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08" containerID="1667e55d178e8f38a692c5ea94bde08a6200d42064af14191e9ac90e82e74946" exitCode=0 Dec 03 17:48:55 crc kubenswrapper[4787]: I1203 17:48:55.335366 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" event={"ID":"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08","Type":"ContainerDied","Data":"1667e55d178e8f38a692c5ea94bde08a6200d42064af14191e9ac90e82e74946"} Dec 03 17:48:55 crc kubenswrapper[4787]: I1203 17:48:55.391763 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:56 crc kubenswrapper[4787]: I1203 17:48:56.851285 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:56 crc kubenswrapper[4787]: I1203 17:48:56.879884 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jst8n\" (UniqueName: \"kubernetes.io/projected/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-kube-api-access-jst8n\") pod \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " Dec 03 17:48:56 crc kubenswrapper[4787]: I1203 17:48:56.879961 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-ssh-key\") pod \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " Dec 03 17:48:56 crc kubenswrapper[4787]: I1203 17:48:56.880013 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-inventory\") pod \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\" (UID: \"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08\") " Dec 03 17:48:56 crc kubenswrapper[4787]: I1203 17:48:56.891311 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-kube-api-access-jst8n" (OuterVolumeSpecName: "kube-api-access-jst8n") pod "dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08" (UID: "dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08"). InnerVolumeSpecName "kube-api-access-jst8n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:48:56 crc kubenswrapper[4787]: I1203 17:48:56.912713 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08" (UID: "dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:48:56 crc kubenswrapper[4787]: I1203 17:48:56.928377 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-inventory" (OuterVolumeSpecName: "inventory") pod "dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08" (UID: "dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:48:56 crc kubenswrapper[4787]: I1203 17:48:56.984250 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jst8n\" (UniqueName: \"kubernetes.io/projected/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-kube-api-access-jst8n\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:56 crc kubenswrapper[4787]: I1203 17:48:56.984326 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:56 crc kubenswrapper[4787]: I1203 17:48:56.984350 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.361009 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" event={"ID":"dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08","Type":"ContainerDied","Data":"029224a694b97fdeda51ba6f1b5927e0ca46db1ae5c8f483115eb7de16ef08c1"} Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.361455 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="029224a694b97fdeda51ba6f1b5927e0ca46db1ae5c8f483115eb7de16ef08c1" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.361139 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2" Dec 03 17:48:57 crc kubenswrapper[4787]: E1203 17:48:57.465610 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddbaf8a7d_b17e_4a83_a754_cec5ff3d6a08.slice/crio-029224a694b97fdeda51ba6f1b5927e0ca46db1ae5c8f483115eb7de16ef08c1\": RecentStats: unable to find data in memory cache]" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.467476 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn"] Dec 03 17:48:57 crc kubenswrapper[4787]: E1203 17:48:57.468036 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.468058 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.468471 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.469784 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.471818 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.472572 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.472831 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.473502 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.488995 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn"] Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.599590 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dwlg\" (UniqueName: \"kubernetes.io/projected/f34ffd31-6238-4f68-80cc-f07be91a44f8-kube-api-access-5dwlg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.599754 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.599788 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.701994 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dwlg\" (UniqueName: \"kubernetes.io/projected/f34ffd31-6238-4f68-80cc-f07be91a44f8-kube-api-access-5dwlg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.702201 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.702245 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.707625 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.709240 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.722566 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dwlg\" (UniqueName: \"kubernetes.io/projected/f34ffd31-6238-4f68-80cc-f07be91a44f8-kube-api-access-5dwlg\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:48:57 crc kubenswrapper[4787]: I1203 17:48:57.794908 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:48:58 crc kubenswrapper[4787]: I1203 17:48:58.139812 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-29l7v"] Dec 03 17:48:58 crc kubenswrapper[4787]: I1203 17:48:58.327244 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn"] Dec 03 17:48:58 crc kubenswrapper[4787]: I1203 17:48:58.371282 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" event={"ID":"f34ffd31-6238-4f68-80cc-f07be91a44f8","Type":"ContainerStarted","Data":"d0e999d0b289d97c587d44616f0872403163ecfdd5f308c68a05c947c249b4b4"} Dec 03 17:48:58 crc kubenswrapper[4787]: I1203 17:48:58.371481 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-29l7v" podUID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" containerName="registry-server" containerID="cri-o://25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a" gracePeriod=2 Dec 03 17:48:58 crc kubenswrapper[4787]: I1203 17:48:58.960801 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.034156 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-utilities\") pod \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.034216 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-catalog-content\") pod \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.034523 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpmvs\" (UniqueName: \"kubernetes.io/projected/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-kube-api-access-bpmvs\") pod \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\" (UID: \"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb\") " Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.035312 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-utilities" (OuterVolumeSpecName: "utilities") pod "e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" (UID: "e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.047455 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-kube-api-access-bpmvs" (OuterVolumeSpecName: "kube-api-access-bpmvs") pod "e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" (UID: "e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb"). InnerVolumeSpecName "kube-api-access-bpmvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.054423 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" (UID: "e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.136985 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpmvs\" (UniqueName: \"kubernetes.io/projected/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-kube-api-access-bpmvs\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.137051 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.137061 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.392263 4787 generic.go:334] "Generic (PLEG): container finished" podID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" containerID="25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a" exitCode=0 Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.392347 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-29l7v" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.392344 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29l7v" event={"ID":"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb","Type":"ContainerDied","Data":"25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a"} Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.392490 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-29l7v" event={"ID":"e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb","Type":"ContainerDied","Data":"41739f6c8cdff8579bed569af1864b8a956761b352674003b34940457e8731d9"} Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.392515 4787 scope.go:117] "RemoveContainer" containerID="25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.396428 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" event={"ID":"f34ffd31-6238-4f68-80cc-f07be91a44f8","Type":"ContainerStarted","Data":"74b0cc42250f3c808d019102b9ee07c7623bbfd7ac34b6181b9ff01c02036a9e"} Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.427689 4787 scope.go:117] "RemoveContainer" containerID="bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.431279 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" podStartSLOduration=1.958786002 podStartE2EDuration="2.431252382s" podCreationTimestamp="2025-12-03 17:48:57 +0000 UTC" firstStartedPulling="2025-12-03 17:48:58.329730259 +0000 UTC m=+2155.147201218" lastFinishedPulling="2025-12-03 17:48:58.802196639 +0000 UTC m=+2155.619667598" observedRunningTime="2025-12-03 17:48:59.41404299 +0000 UTC m=+2156.231513949" watchObservedRunningTime="2025-12-03 17:48:59.431252382 +0000 UTC m=+2156.248723341" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.447024 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-29l7v"] Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.452896 4787 scope.go:117] "RemoveContainer" containerID="1f7f4253b84502eefb3103e56a8189f82f34c96a8a612eca85441deec62152ed" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.454715 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-29l7v"] Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.505518 4787 scope.go:117] "RemoveContainer" containerID="25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a" Dec 03 17:48:59 crc kubenswrapper[4787]: E1203 17:48:59.506003 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a\": container with ID starting with 25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a not found: ID does not exist" containerID="25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.506066 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a"} err="failed to get container status \"25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a\": rpc error: code = NotFound desc = could not find container \"25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a\": container with ID starting with 25a0b7313921ff16a19f20582a3c963a053c5d3619245160bf19e6b84fa7aa3a not found: ID does not exist" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.506098 4787 scope.go:117] "RemoveContainer" containerID="bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223" Dec 03 17:48:59 crc kubenswrapper[4787]: E1203 17:48:59.506492 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223\": container with ID starting with bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223 not found: ID does not exist" containerID="bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.506524 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223"} err="failed to get container status \"bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223\": rpc error: code = NotFound desc = could not find container \"bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223\": container with ID starting with bbab943a65d0c1871afc16e809faa0d4cee76b475a1bbb7cb43e8eab8af83223 not found: ID does not exist" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.506549 4787 scope.go:117] "RemoveContainer" containerID="1f7f4253b84502eefb3103e56a8189f82f34c96a8a612eca85441deec62152ed" Dec 03 17:48:59 crc kubenswrapper[4787]: E1203 17:48:59.506865 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f7f4253b84502eefb3103e56a8189f82f34c96a8a612eca85441deec62152ed\": container with ID starting with 1f7f4253b84502eefb3103e56a8189f82f34c96a8a612eca85441deec62152ed not found: ID does not exist" containerID="1f7f4253b84502eefb3103e56a8189f82f34c96a8a612eca85441deec62152ed" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.506977 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f7f4253b84502eefb3103e56a8189f82f34c96a8a612eca85441deec62152ed"} err="failed to get container status \"1f7f4253b84502eefb3103e56a8189f82f34c96a8a612eca85441deec62152ed\": rpc error: code = NotFound desc = could not find container \"1f7f4253b84502eefb3103e56a8189f82f34c96a8a612eca85441deec62152ed\": container with ID starting with 1f7f4253b84502eefb3103e56a8189f82f34c96a8a612eca85441deec62152ed not found: ID does not exist" Dec 03 17:48:59 crc kubenswrapper[4787]: I1203 17:48:59.781920 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" path="/var/lib/kubelet/pods/e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb/volumes" Dec 03 17:49:09 crc kubenswrapper[4787]: I1203 17:49:09.510098 4787 generic.go:334] "Generic (PLEG): container finished" podID="f34ffd31-6238-4f68-80cc-f07be91a44f8" containerID="74b0cc42250f3c808d019102b9ee07c7623bbfd7ac34b6181b9ff01c02036a9e" exitCode=0 Dec 03 17:49:09 crc kubenswrapper[4787]: I1203 17:49:09.510626 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" event={"ID":"f34ffd31-6238-4f68-80cc-f07be91a44f8","Type":"ContainerDied","Data":"74b0cc42250f3c808d019102b9ee07c7623bbfd7ac34b6181b9ff01c02036a9e"} Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.032500 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.189438 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5dwlg\" (UniqueName: \"kubernetes.io/projected/f34ffd31-6238-4f68-80cc-f07be91a44f8-kube-api-access-5dwlg\") pod \"f34ffd31-6238-4f68-80cc-f07be91a44f8\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.189550 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-ssh-key\") pod \"f34ffd31-6238-4f68-80cc-f07be91a44f8\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.189714 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-inventory\") pod \"f34ffd31-6238-4f68-80cc-f07be91a44f8\" (UID: \"f34ffd31-6238-4f68-80cc-f07be91a44f8\") " Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.195488 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f34ffd31-6238-4f68-80cc-f07be91a44f8-kube-api-access-5dwlg" (OuterVolumeSpecName: "kube-api-access-5dwlg") pod "f34ffd31-6238-4f68-80cc-f07be91a44f8" (UID: "f34ffd31-6238-4f68-80cc-f07be91a44f8"). InnerVolumeSpecName "kube-api-access-5dwlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.224849 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f34ffd31-6238-4f68-80cc-f07be91a44f8" (UID: "f34ffd31-6238-4f68-80cc-f07be91a44f8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.225508 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-inventory" (OuterVolumeSpecName: "inventory") pod "f34ffd31-6238-4f68-80cc-f07be91a44f8" (UID: "f34ffd31-6238-4f68-80cc-f07be91a44f8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.293253 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.293692 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5dwlg\" (UniqueName: \"kubernetes.io/projected/f34ffd31-6238-4f68-80cc-f07be91a44f8-kube-api-access-5dwlg\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.293716 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f34ffd31-6238-4f68-80cc-f07be91a44f8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.531001 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" event={"ID":"f34ffd31-6238-4f68-80cc-f07be91a44f8","Type":"ContainerDied","Data":"d0e999d0b289d97c587d44616f0872403163ecfdd5f308c68a05c947c249b4b4"} Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.531071 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d0e999d0b289d97c587d44616f0872403163ecfdd5f308c68a05c947c249b4b4" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.531130 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.605972 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f"] Dec 03 17:49:11 crc kubenswrapper[4787]: E1203 17:49:11.606511 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" containerName="registry-server" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.606533 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" containerName="registry-server" Dec 03 17:49:11 crc kubenswrapper[4787]: E1203 17:49:11.606562 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" containerName="extract-content" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.606571 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" containerName="extract-content" Dec 03 17:49:11 crc kubenswrapper[4787]: E1203 17:49:11.606594 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f34ffd31-6238-4f68-80cc-f07be91a44f8" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.606605 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f34ffd31-6238-4f68-80cc-f07be91a44f8" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:49:11 crc kubenswrapper[4787]: E1203 17:49:11.606629 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" containerName="extract-utilities" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.606638 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" containerName="extract-utilities" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.606925 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e48ce5f7-6436-4490-a0f4-8d80bdc2c6eb" containerName="registry-server" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.606996 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f34ffd31-6238-4f68-80cc-f07be91a44f8" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.607837 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.610704 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.611121 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.611243 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.611369 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.612106 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.612230 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.613824 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.624088 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.625419 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f"] Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.804424 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.804729 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zh6r\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-kube-api-access-4zh6r\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.804821 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.804905 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.805073 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.805191 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.805277 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.805372 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.805446 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.805530 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.805623 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.805823 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.806006 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.908643 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.909061 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.909334 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.910645 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zh6r\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-kube-api-access-4zh6r\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.910753 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.910860 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.911059 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.911178 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.911283 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.911405 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.911506 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.911617 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.911741 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.921781 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.921931 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.927716 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.928304 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.932879 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.935727 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.935825 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.936382 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.937376 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.937579 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.937732 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.942749 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zh6r\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-kube-api-access-4zh6r\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:11 crc kubenswrapper[4787]: I1203 17:49:11.943665 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-dql8f\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:12 crc kubenswrapper[4787]: I1203 17:49:12.226782 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:12 crc kubenswrapper[4787]: I1203 17:49:12.772395 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f"] Dec 03 17:49:13 crc kubenswrapper[4787]: I1203 17:49:13.053365 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-7gzqf"] Dec 03 17:49:13 crc kubenswrapper[4787]: I1203 17:49:13.065102 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-7gzqf"] Dec 03 17:49:13 crc kubenswrapper[4787]: I1203 17:49:13.554770 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" event={"ID":"c408e255-873c-4200-9768-6576ea6f6b0c","Type":"ContainerStarted","Data":"c1417c44ed98bbcdd04a1d14e588aaa4a1a70f862af5970056a9ec4c18519b23"} Dec 03 17:49:13 crc kubenswrapper[4787]: I1203 17:49:13.787190 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a28737aa-7ebe-4dbe-967d-51376202958e" path="/var/lib/kubelet/pods/a28737aa-7ebe-4dbe-967d-51376202958e/volumes" Dec 03 17:49:14 crc kubenswrapper[4787]: I1203 17:49:14.566094 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" event={"ID":"c408e255-873c-4200-9768-6576ea6f6b0c","Type":"ContainerStarted","Data":"da01f1d732c8465b9ef4e94f3c6efd89b90a966451ccf642ee3263a38db1e7d8"} Dec 03 17:49:14 crc kubenswrapper[4787]: I1203 17:49:14.585128 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" podStartSLOduration=3.039687755 podStartE2EDuration="3.585107713s" podCreationTimestamp="2025-12-03 17:49:11 +0000 UTC" firstStartedPulling="2025-12-03 17:49:12.775270302 +0000 UTC m=+2169.592741261" lastFinishedPulling="2025-12-03 17:49:13.32069026 +0000 UTC m=+2170.138161219" observedRunningTime="2025-12-03 17:49:14.583250523 +0000 UTC m=+2171.400721482" watchObservedRunningTime="2025-12-03 17:49:14.585107713 +0000 UTC m=+2171.402578672" Dec 03 17:49:18 crc kubenswrapper[4787]: I1203 17:49:18.989570 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:49:18 crc kubenswrapper[4787]: I1203 17:49:18.990174 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:49:24 crc kubenswrapper[4787]: I1203 17:49:24.563632 4787 scope.go:117] "RemoveContainer" containerID="1d5521e67009ea94ea1bdda51184a2855b1c1209b56b5730ab6af30fca5119e6" Dec 03 17:49:24 crc kubenswrapper[4787]: I1203 17:49:24.627488 4787 scope.go:117] "RemoveContainer" containerID="1a00a16219839ac3284d87eacc77080bb5386e413634aaad555e7c56871e39f0" Dec 03 17:49:24 crc kubenswrapper[4787]: I1203 17:49:24.673628 4787 scope.go:117] "RemoveContainer" containerID="044d86c2a2fe3dbd8a3c6756e51f420f54addb310c01feaa7100fba82c302de5" Dec 03 17:49:48 crc kubenswrapper[4787]: I1203 17:49:48.990440 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:49:48 crc kubenswrapper[4787]: I1203 17:49:48.991055 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:49:48 crc kubenswrapper[4787]: I1203 17:49:48.991110 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:49:48 crc kubenswrapper[4787]: I1203 17:49:48.991758 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c2571bed769c5e81a57d85c403092e3b76ad0831c285bb44f058f97347d8398e"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:49:48 crc kubenswrapper[4787]: I1203 17:49:48.991824 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://c2571bed769c5e81a57d85c403092e3b76ad0831c285bb44f058f97347d8398e" gracePeriod=600 Dec 03 17:49:49 crc kubenswrapper[4787]: I1203 17:49:49.933653 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="c2571bed769c5e81a57d85c403092e3b76ad0831c285bb44f058f97347d8398e" exitCode=0 Dec 03 17:49:49 crc kubenswrapper[4787]: I1203 17:49:49.933734 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"c2571bed769c5e81a57d85c403092e3b76ad0831c285bb44f058f97347d8398e"} Dec 03 17:49:49 crc kubenswrapper[4787]: I1203 17:49:49.934180 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9"} Dec 03 17:49:49 crc kubenswrapper[4787]: I1203 17:49:49.934199 4787 scope.go:117] "RemoveContainer" containerID="8c50dc4bbf01f3531929a156d2fdc24487e8f19415de29e3c7a68d792797af75" Dec 03 17:49:50 crc kubenswrapper[4787]: I1203 17:49:50.947303 4787 generic.go:334] "Generic (PLEG): container finished" podID="c408e255-873c-4200-9768-6576ea6f6b0c" containerID="da01f1d732c8465b9ef4e94f3c6efd89b90a966451ccf642ee3263a38db1e7d8" exitCode=0 Dec 03 17:49:50 crc kubenswrapper[4787]: I1203 17:49:50.947501 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" event={"ID":"c408e255-873c-4200-9768-6576ea6f6b0c","Type":"ContainerDied","Data":"da01f1d732c8465b9ef4e94f3c6efd89b90a966451ccf642ee3263a38db1e7d8"} Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.498452 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564307 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564360 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-inventory\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564448 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564549 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-bootstrap-combined-ca-bundle\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564606 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-repo-setup-combined-ca-bundle\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564665 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-ovn-default-certs-0\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564735 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ovn-combined-ca-bundle\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564773 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-combined-ca-bundle\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564801 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564896 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-power-monitoring-combined-ca-bundle\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564935 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-libvirt-combined-ca-bundle\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.564970 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ssh-key\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.565009 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zh6r\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-kube-api-access-4zh6r\") pod \"c408e255-873c-4200-9768-6576ea6f6b0c\" (UID: \"c408e255-873c-4200-9768-6576ea6f6b0c\") " Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.572938 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.572985 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-kube-api-access-4zh6r" (OuterVolumeSpecName: "kube-api-access-4zh6r") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "kube-api-access-4zh6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.574631 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.575659 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.576965 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.579935 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.579955 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.580979 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.581367 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.581808 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.584381 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.614284 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-inventory" (OuterVolumeSpecName: "inventory") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.619147 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c408e255-873c-4200-9768-6576ea6f6b0c" (UID: "c408e255-873c-4200-9768-6576ea6f6b0c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667738 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667773 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667784 4787 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667794 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667805 4787 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667816 4787 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667824 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667832 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zh6r\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-kube-api-access-4zh6r\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667842 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667851 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667860 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/c408e255-873c-4200-9768-6576ea6f6b0c-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667869 4787 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.667877 4787 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c408e255-873c-4200-9768-6576ea6f6b0c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.970573 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" event={"ID":"c408e255-873c-4200-9768-6576ea6f6b0c","Type":"ContainerDied","Data":"c1417c44ed98bbcdd04a1d14e588aaa4a1a70f862af5970056a9ec4c18519b23"} Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.970617 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c1417c44ed98bbcdd04a1d14e588aaa4a1a70f862af5970056a9ec4c18519b23" Dec 03 17:49:52 crc kubenswrapper[4787]: I1203 17:49:52.970639 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.096572 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd"] Dec 03 17:49:53 crc kubenswrapper[4787]: E1203 17:49:53.097231 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c408e255-873c-4200-9768-6576ea6f6b0c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.097265 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c408e255-873c-4200-9768-6576ea6f6b0c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.097580 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c408e255-873c-4200-9768-6576ea6f6b0c" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.098723 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.107065 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.107319 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.107565 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.107964 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.108180 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.111174 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd"] Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.177826 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.178076 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.178162 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.178180 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.178200 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8zf9l\" (UniqueName: \"kubernetes.io/projected/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-kube-api-access-8zf9l\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.280348 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.280457 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.280560 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.280583 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.280609 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8zf9l\" (UniqueName: \"kubernetes.io/projected/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-kube-api-access-8zf9l\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.281339 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.286070 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.286572 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.290432 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.299078 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8zf9l\" (UniqueName: \"kubernetes.io/projected/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-kube-api-access-8zf9l\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-s4chd\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:53 crc kubenswrapper[4787]: I1203 17:49:53.437001 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:49:54 crc kubenswrapper[4787]: I1203 17:49:54.021500 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd"] Dec 03 17:49:54 crc kubenswrapper[4787]: W1203 17:49:54.023230 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5d4c7463_edc8_40f9_80f6_65fe7d4e40b8.slice/crio-f036b64369178449273378d4af4498ea2d1eae489db48f27fdb1e10c2338ad28 WatchSource:0}: Error finding container f036b64369178449273378d4af4498ea2d1eae489db48f27fdb1e10c2338ad28: Status 404 returned error can't find the container with id f036b64369178449273378d4af4498ea2d1eae489db48f27fdb1e10c2338ad28 Dec 03 17:49:54 crc kubenswrapper[4787]: I1203 17:49:54.997559 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" event={"ID":"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8","Type":"ContainerStarted","Data":"9365248c1aca69a164c2213cd1d1583fdcf9f1b00170793994c6a824c715ccad"} Dec 03 17:49:54 crc kubenswrapper[4787]: I1203 17:49:54.998108 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" event={"ID":"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8","Type":"ContainerStarted","Data":"f036b64369178449273378d4af4498ea2d1eae489db48f27fdb1e10c2338ad28"} Dec 03 17:49:55 crc kubenswrapper[4787]: I1203 17:49:55.031507 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" podStartSLOduration=1.484886131 podStartE2EDuration="2.031486663s" podCreationTimestamp="2025-12-03 17:49:53 +0000 UTC" firstStartedPulling="2025-12-03 17:49:54.026644864 +0000 UTC m=+2210.844115823" lastFinishedPulling="2025-12-03 17:49:54.573245396 +0000 UTC m=+2211.390716355" observedRunningTime="2025-12-03 17:49:55.017511268 +0000 UTC m=+2211.834982247" watchObservedRunningTime="2025-12-03 17:49:55.031486663 +0000 UTC m=+2211.848957632" Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.470110 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ctvvj"] Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.472996 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.487663 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ctvvj"] Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.654294 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-catalog-content\") pod \"community-operators-ctvvj\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.654377 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-utilities\") pod \"community-operators-ctvvj\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.654601 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g72fn\" (UniqueName: \"kubernetes.io/projected/e2f94a8e-6e7f-447c-9853-c94187404712-kube-api-access-g72fn\") pod \"community-operators-ctvvj\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.757350 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-utilities\") pod \"community-operators-ctvvj\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.758098 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-utilities\") pod \"community-operators-ctvvj\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.758120 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g72fn\" (UniqueName: \"kubernetes.io/projected/e2f94a8e-6e7f-447c-9853-c94187404712-kube-api-access-g72fn\") pod \"community-operators-ctvvj\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.758246 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-catalog-content\") pod \"community-operators-ctvvj\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.758589 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-catalog-content\") pod \"community-operators-ctvvj\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.787688 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g72fn\" (UniqueName: \"kubernetes.io/projected/e2f94a8e-6e7f-447c-9853-c94187404712-kube-api-access-g72fn\") pod \"community-operators-ctvvj\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:09 crc kubenswrapper[4787]: I1203 17:50:09.797749 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:10 crc kubenswrapper[4787]: I1203 17:50:10.522495 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ctvvj"] Dec 03 17:50:10 crc kubenswrapper[4787]: W1203 17:50:10.526327 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2f94a8e_6e7f_447c_9853_c94187404712.slice/crio-e1657ec6ece25f1c4b1afe2a1574af3e39487ace9fb04d699361047b10c17f4f WatchSource:0}: Error finding container e1657ec6ece25f1c4b1afe2a1574af3e39487ace9fb04d699361047b10c17f4f: Status 404 returned error can't find the container with id e1657ec6ece25f1c4b1afe2a1574af3e39487ace9fb04d699361047b10c17f4f Dec 03 17:50:11 crc kubenswrapper[4787]: I1203 17:50:11.214198 4787 generic.go:334] "Generic (PLEG): container finished" podID="e2f94a8e-6e7f-447c-9853-c94187404712" containerID="44566fe3da36293eceda74dfb4d52f757716d2429582c7b2fac94ba939178e87" exitCode=0 Dec 03 17:50:11 crc kubenswrapper[4787]: I1203 17:50:11.214410 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ctvvj" event={"ID":"e2f94a8e-6e7f-447c-9853-c94187404712","Type":"ContainerDied","Data":"44566fe3da36293eceda74dfb4d52f757716d2429582c7b2fac94ba939178e87"} Dec 03 17:50:11 crc kubenswrapper[4787]: I1203 17:50:11.214736 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ctvvj" event={"ID":"e2f94a8e-6e7f-447c-9853-c94187404712","Type":"ContainerStarted","Data":"e1657ec6ece25f1c4b1afe2a1574af3e39487ace9fb04d699361047b10c17f4f"} Dec 03 17:50:12 crc kubenswrapper[4787]: I1203 17:50:12.230670 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ctvvj" event={"ID":"e2f94a8e-6e7f-447c-9853-c94187404712","Type":"ContainerStarted","Data":"d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919"} Dec 03 17:50:13 crc kubenswrapper[4787]: I1203 17:50:13.242853 4787 generic.go:334] "Generic (PLEG): container finished" podID="e2f94a8e-6e7f-447c-9853-c94187404712" containerID="d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919" exitCode=0 Dec 03 17:50:13 crc kubenswrapper[4787]: I1203 17:50:13.242951 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ctvvj" event={"ID":"e2f94a8e-6e7f-447c-9853-c94187404712","Type":"ContainerDied","Data":"d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919"} Dec 03 17:50:14 crc kubenswrapper[4787]: I1203 17:50:14.254897 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ctvvj" event={"ID":"e2f94a8e-6e7f-447c-9853-c94187404712","Type":"ContainerStarted","Data":"9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469"} Dec 03 17:50:14 crc kubenswrapper[4787]: I1203 17:50:14.283848 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ctvvj" podStartSLOduration=2.849116868 podStartE2EDuration="5.283823991s" podCreationTimestamp="2025-12-03 17:50:09 +0000 UTC" firstStartedPulling="2025-12-03 17:50:11.216417532 +0000 UTC m=+2228.033888531" lastFinishedPulling="2025-12-03 17:50:13.651124695 +0000 UTC m=+2230.468595654" observedRunningTime="2025-12-03 17:50:14.276262648 +0000 UTC m=+2231.093733607" watchObservedRunningTime="2025-12-03 17:50:14.283823991 +0000 UTC m=+2231.101294950" Dec 03 17:50:19 crc kubenswrapper[4787]: I1203 17:50:19.798493 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:19 crc kubenswrapper[4787]: I1203 17:50:19.799204 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:19 crc kubenswrapper[4787]: I1203 17:50:19.881064 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:20 crc kubenswrapper[4787]: I1203 17:50:20.389707 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:20 crc kubenswrapper[4787]: I1203 17:50:20.455233 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ctvvj"] Dec 03 17:50:22 crc kubenswrapper[4787]: I1203 17:50:22.340955 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ctvvj" podUID="e2f94a8e-6e7f-447c-9853-c94187404712" containerName="registry-server" containerID="cri-o://9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469" gracePeriod=2 Dec 03 17:50:22 crc kubenswrapper[4787]: I1203 17:50:22.872343 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.055895 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-catalog-content\") pod \"e2f94a8e-6e7f-447c-9853-c94187404712\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.056059 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g72fn\" (UniqueName: \"kubernetes.io/projected/e2f94a8e-6e7f-447c-9853-c94187404712-kube-api-access-g72fn\") pod \"e2f94a8e-6e7f-447c-9853-c94187404712\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.056417 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-utilities\") pod \"e2f94a8e-6e7f-447c-9853-c94187404712\" (UID: \"e2f94a8e-6e7f-447c-9853-c94187404712\") " Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.057182 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-utilities" (OuterVolumeSpecName: "utilities") pod "e2f94a8e-6e7f-447c-9853-c94187404712" (UID: "e2f94a8e-6e7f-447c-9853-c94187404712"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.065123 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2f94a8e-6e7f-447c-9853-c94187404712-kube-api-access-g72fn" (OuterVolumeSpecName: "kube-api-access-g72fn") pod "e2f94a8e-6e7f-447c-9853-c94187404712" (UID: "e2f94a8e-6e7f-447c-9853-c94187404712"). InnerVolumeSpecName "kube-api-access-g72fn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.113819 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e2f94a8e-6e7f-447c-9853-c94187404712" (UID: "e2f94a8e-6e7f-447c-9853-c94187404712"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.159447 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.159763 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e2f94a8e-6e7f-447c-9853-c94187404712-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.159882 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g72fn\" (UniqueName: \"kubernetes.io/projected/e2f94a8e-6e7f-447c-9853-c94187404712-kube-api-access-g72fn\") on node \"crc\" DevicePath \"\"" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.360047 4787 generic.go:334] "Generic (PLEG): container finished" podID="e2f94a8e-6e7f-447c-9853-c94187404712" containerID="9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469" exitCode=0 Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.360094 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ctvvj" event={"ID":"e2f94a8e-6e7f-447c-9853-c94187404712","Type":"ContainerDied","Data":"9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469"} Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.360123 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ctvvj" event={"ID":"e2f94a8e-6e7f-447c-9853-c94187404712","Type":"ContainerDied","Data":"e1657ec6ece25f1c4b1afe2a1574af3e39487ace9fb04d699361047b10c17f4f"} Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.360141 4787 scope.go:117] "RemoveContainer" containerID="9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.360178 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ctvvj" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.389462 4787 scope.go:117] "RemoveContainer" containerID="d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.402670 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ctvvj"] Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.414030 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ctvvj"] Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.422466 4787 scope.go:117] "RemoveContainer" containerID="44566fe3da36293eceda74dfb4d52f757716d2429582c7b2fac94ba939178e87" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.484490 4787 scope.go:117] "RemoveContainer" containerID="9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469" Dec 03 17:50:23 crc kubenswrapper[4787]: E1203 17:50:23.485078 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469\": container with ID starting with 9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469 not found: ID does not exist" containerID="9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.485133 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469"} err="failed to get container status \"9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469\": rpc error: code = NotFound desc = could not find container \"9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469\": container with ID starting with 9d2d3f72794e8f9b260f2b49674f4a53698cbb0c47f40dc22907589a2e8a1469 not found: ID does not exist" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.485171 4787 scope.go:117] "RemoveContainer" containerID="d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919" Dec 03 17:50:23 crc kubenswrapper[4787]: E1203 17:50:23.485736 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919\": container with ID starting with d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919 not found: ID does not exist" containerID="d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.485779 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919"} err="failed to get container status \"d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919\": rpc error: code = NotFound desc = could not find container \"d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919\": container with ID starting with d4b7e0b4b84f821107ad9ac0fb9b585bd168b5385843510c3ce4e1a62b48f919 not found: ID does not exist" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.485805 4787 scope.go:117] "RemoveContainer" containerID="44566fe3da36293eceda74dfb4d52f757716d2429582c7b2fac94ba939178e87" Dec 03 17:50:23 crc kubenswrapper[4787]: E1203 17:50:23.486198 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"44566fe3da36293eceda74dfb4d52f757716d2429582c7b2fac94ba939178e87\": container with ID starting with 44566fe3da36293eceda74dfb4d52f757716d2429582c7b2fac94ba939178e87 not found: ID does not exist" containerID="44566fe3da36293eceda74dfb4d52f757716d2429582c7b2fac94ba939178e87" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.486222 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"44566fe3da36293eceda74dfb4d52f757716d2429582c7b2fac94ba939178e87"} err="failed to get container status \"44566fe3da36293eceda74dfb4d52f757716d2429582c7b2fac94ba939178e87\": rpc error: code = NotFound desc = could not find container \"44566fe3da36293eceda74dfb4d52f757716d2429582c7b2fac94ba939178e87\": container with ID starting with 44566fe3da36293eceda74dfb4d52f757716d2429582c7b2fac94ba939178e87 not found: ID does not exist" Dec 03 17:50:23 crc kubenswrapper[4787]: I1203 17:50:23.779634 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2f94a8e-6e7f-447c-9853-c94187404712" path="/var/lib/kubelet/pods/e2f94a8e-6e7f-447c-9853-c94187404712/volumes" Dec 03 17:51:06 crc kubenswrapper[4787]: I1203 17:51:06.838462 4787 generic.go:334] "Generic (PLEG): container finished" podID="5d4c7463-edc8-40f9-80f6-65fe7d4e40b8" containerID="9365248c1aca69a164c2213cd1d1583fdcf9f1b00170793994c6a824c715ccad" exitCode=0 Dec 03 17:51:06 crc kubenswrapper[4787]: I1203 17:51:06.838553 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" event={"ID":"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8","Type":"ContainerDied","Data":"9365248c1aca69a164c2213cd1d1583fdcf9f1b00170793994c6a824c715ccad"} Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.330222 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.472657 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovncontroller-config-0\") pod \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.472731 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8zf9l\" (UniqueName: \"kubernetes.io/projected/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-kube-api-access-8zf9l\") pod \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.472807 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovn-combined-ca-bundle\") pod \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.472847 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ssh-key\") pod \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.472892 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-inventory\") pod \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\" (UID: \"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8\") " Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.479045 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-kube-api-access-8zf9l" (OuterVolumeSpecName: "kube-api-access-8zf9l") pod "5d4c7463-edc8-40f9-80f6-65fe7d4e40b8" (UID: "5d4c7463-edc8-40f9-80f6-65fe7d4e40b8"). InnerVolumeSpecName "kube-api-access-8zf9l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.480330 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "5d4c7463-edc8-40f9-80f6-65fe7d4e40b8" (UID: "5d4c7463-edc8-40f9-80f6-65fe7d4e40b8"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.501693 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "5d4c7463-edc8-40f9-80f6-65fe7d4e40b8" (UID: "5d4c7463-edc8-40f9-80f6-65fe7d4e40b8"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.504179 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5d4c7463-edc8-40f9-80f6-65fe7d4e40b8" (UID: "5d4c7463-edc8-40f9-80f6-65fe7d4e40b8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.535250 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-inventory" (OuterVolumeSpecName: "inventory") pod "5d4c7463-edc8-40f9-80f6-65fe7d4e40b8" (UID: "5d4c7463-edc8-40f9-80f6-65fe7d4e40b8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.575823 4787 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.575860 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8zf9l\" (UniqueName: \"kubernetes.io/projected/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-kube-api-access-8zf9l\") on node \"crc\" DevicePath \"\"" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.575869 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.575878 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.575887 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.862130 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" event={"ID":"5d4c7463-edc8-40f9-80f6-65fe7d4e40b8","Type":"ContainerDied","Data":"f036b64369178449273378d4af4498ea2d1eae489db48f27fdb1e10c2338ad28"} Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.862194 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f036b64369178449273378d4af4498ea2d1eae489db48f27fdb1e10c2338ad28" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.862217 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.977553 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g"] Dec 03 17:51:08 crc kubenswrapper[4787]: E1203 17:51:08.978304 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2f94a8e-6e7f-447c-9853-c94187404712" containerName="extract-content" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.978339 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2f94a8e-6e7f-447c-9853-c94187404712" containerName="extract-content" Dec 03 17:51:08 crc kubenswrapper[4787]: E1203 17:51:08.978393 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2f94a8e-6e7f-447c-9853-c94187404712" containerName="registry-server" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.978407 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2f94a8e-6e7f-447c-9853-c94187404712" containerName="registry-server" Dec 03 17:51:08 crc kubenswrapper[4787]: E1203 17:51:08.978452 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d4c7463-edc8-40f9-80f6-65fe7d4e40b8" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.978465 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d4c7463-edc8-40f9-80f6-65fe7d4e40b8" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 17:51:08 crc kubenswrapper[4787]: E1203 17:51:08.978486 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2f94a8e-6e7f-447c-9853-c94187404712" containerName="extract-utilities" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.978498 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2f94a8e-6e7f-447c-9853-c94187404712" containerName="extract-utilities" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.978816 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2f94a8e-6e7f-447c-9853-c94187404712" containerName="registry-server" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.978855 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d4c7463-edc8-40f9-80f6-65fe7d4e40b8" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.979874 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.984355 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.984376 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.984268 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.984368 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.984605 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 17:51:08 crc kubenswrapper[4787]: I1203 17:51:08.988179 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g"] Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.090920 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.090976 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.091118 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.091255 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.091755 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h6btg\" (UniqueName: \"kubernetes.io/projected/251ef657-822b-4392-9c92-266ccd5ca952-kube-api-access-h6btg\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.193501 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.193622 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.193825 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h6btg\" (UniqueName: \"kubernetes.io/projected/251ef657-822b-4392-9c92-266ccd5ca952-kube-api-access-h6btg\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.194274 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.194734 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.197847 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.197851 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.199164 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.199235 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.212632 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h6btg\" (UniqueName: \"kubernetes.io/projected/251ef657-822b-4392-9c92-266ccd5ca952-kube-api-access-h6btg\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.319143 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:51:09 crc kubenswrapper[4787]: I1203 17:51:09.905002 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g"] Dec 03 17:51:10 crc kubenswrapper[4787]: I1203 17:51:10.885061 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" event={"ID":"251ef657-822b-4392-9c92-266ccd5ca952","Type":"ContainerStarted","Data":"762d7ae422f90cd7ab8b2856d50541644f647d79c9c4d20957c45e0fbf9075ef"} Dec 03 17:51:10 crc kubenswrapper[4787]: I1203 17:51:10.885811 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" event={"ID":"251ef657-822b-4392-9c92-266ccd5ca952","Type":"ContainerStarted","Data":"ee7051c0e9ed4c805082e13a6b6aa165879955e750f78578fdd3af8612940059"} Dec 03 17:51:10 crc kubenswrapper[4787]: I1203 17:51:10.901117 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" podStartSLOduration=2.4860931219999998 podStartE2EDuration="2.901099278s" podCreationTimestamp="2025-12-03 17:51:08 +0000 UTC" firstStartedPulling="2025-12-03 17:51:09.918724312 +0000 UTC m=+2286.736195271" lastFinishedPulling="2025-12-03 17:51:10.333730458 +0000 UTC m=+2287.151201427" observedRunningTime="2025-12-03 17:51:10.900355358 +0000 UTC m=+2287.717826337" watchObservedRunningTime="2025-12-03 17:51:10.901099278 +0000 UTC m=+2287.718570237" Dec 03 17:51:12 crc kubenswrapper[4787]: I1203 17:51:12.041010 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/heat-db-sync-2gfq8"] Dec 03 17:51:12 crc kubenswrapper[4787]: I1203 17:51:12.054530 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/heat-db-sync-2gfq8"] Dec 03 17:51:13 crc kubenswrapper[4787]: I1203 17:51:13.786594 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5749d651-999d-4d54-8661-1962fc41521c" path="/var/lib/kubelet/pods/5749d651-999d-4d54-8661-1962fc41521c/volumes" Dec 03 17:51:24 crc kubenswrapper[4787]: I1203 17:51:24.856300 4787 scope.go:117] "RemoveContainer" containerID="beed253452ae9872609aaefd211c3936f663e0f5d08560a30440cbccbe3c7e20" Dec 03 17:51:54 crc kubenswrapper[4787]: I1203 17:51:54.047287 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/aodh-db-sync-62rbm"] Dec 03 17:51:54 crc kubenswrapper[4787]: I1203 17:51:54.056494 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/aodh-db-sync-62rbm"] Dec 03 17:51:55 crc kubenswrapper[4787]: I1203 17:51:55.794228 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68706b1c-146b-4c74-a8b0-4c24ff3758da" path="/var/lib/kubelet/pods/68706b1c-146b-4c74-a8b0-4c24ff3758da/volumes" Dec 03 17:52:18 crc kubenswrapper[4787]: I1203 17:52:18.990220 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:52:18 crc kubenswrapper[4787]: I1203 17:52:18.990925 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:52:24 crc kubenswrapper[4787]: I1203 17:52:24.952402 4787 scope.go:117] "RemoveContainer" containerID="ab006eeba764bb3e104337c95c5ba176b4a6b3ab5a4a144b7be08b5338f6b9dc" Dec 03 17:52:48 crc kubenswrapper[4787]: I1203 17:52:48.990202 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:52:48 crc kubenswrapper[4787]: I1203 17:52:48.990842 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:53:18 crc kubenswrapper[4787]: I1203 17:53:17.872187 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="cert-manager/cert-manager-webhook-5655c58dd6-f46ld" podUID="f5b1552f-1b75-4000-b31e-4d247ea51a65" containerName="cert-manager-webhook" probeResult="failure" output="Get \"http://10.217.0.45:6080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 17:53:18 crc kubenswrapper[4787]: I1203 17:53:18.990038 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 17:53:18 crc kubenswrapper[4787]: I1203 17:53:18.990324 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 17:53:18 crc kubenswrapper[4787]: I1203 17:53:18.990370 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 17:53:18 crc kubenswrapper[4787]: I1203 17:53:18.991137 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 17:53:18 crc kubenswrapper[4787]: I1203 17:53:18.991211 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" gracePeriod=600 Dec 03 17:53:19 crc kubenswrapper[4787]: E1203 17:53:19.115549 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:53:19 crc kubenswrapper[4787]: I1203 17:53:19.446457 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" exitCode=0 Dec 03 17:53:19 crc kubenswrapper[4787]: I1203 17:53:19.446514 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9"} Dec 03 17:53:19 crc kubenswrapper[4787]: I1203 17:53:19.446625 4787 scope.go:117] "RemoveContainer" containerID="c2571bed769c5e81a57d85c403092e3b76ad0831c285bb44f058f97347d8398e" Dec 03 17:53:19 crc kubenswrapper[4787]: I1203 17:53:19.447385 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:53:19 crc kubenswrapper[4787]: E1203 17:53:19.448581 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:53:32 crc kubenswrapper[4787]: I1203 17:53:32.766736 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:53:32 crc kubenswrapper[4787]: E1203 17:53:32.767964 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:53:45 crc kubenswrapper[4787]: I1203 17:53:45.769829 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:53:45 crc kubenswrapper[4787]: E1203 17:53:45.771089 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:53:59 crc kubenswrapper[4787]: I1203 17:53:59.767177 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:53:59 crc kubenswrapper[4787]: E1203 17:53:59.768424 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:54:13 crc kubenswrapper[4787]: I1203 17:54:13.779637 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:54:13 crc kubenswrapper[4787]: E1203 17:54:13.781086 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:54:26 crc kubenswrapper[4787]: I1203 17:54:26.766822 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:54:26 crc kubenswrapper[4787]: E1203 17:54:26.767499 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:54:41 crc kubenswrapper[4787]: I1203 17:54:41.768410 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:54:41 crc kubenswrapper[4787]: E1203 17:54:41.771559 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:54:53 crc kubenswrapper[4787]: I1203 17:54:53.774628 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:54:53 crc kubenswrapper[4787]: E1203 17:54:53.775460 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:55:04 crc kubenswrapper[4787]: I1203 17:55:04.767011 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:55:04 crc kubenswrapper[4787]: E1203 17:55:04.768869 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:55:16 crc kubenswrapper[4787]: I1203 17:55:16.766641 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:55:16 crc kubenswrapper[4787]: E1203 17:55:16.767689 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:55:28 crc kubenswrapper[4787]: I1203 17:55:28.767636 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:55:28 crc kubenswrapper[4787]: E1203 17:55:28.769215 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:55:34 crc kubenswrapper[4787]: I1203 17:55:34.361522 4787 generic.go:334] "Generic (PLEG): container finished" podID="251ef657-822b-4392-9c92-266ccd5ca952" containerID="762d7ae422f90cd7ab8b2856d50541644f647d79c9c4d20957c45e0fbf9075ef" exitCode=0 Dec 03 17:55:34 crc kubenswrapper[4787]: I1203 17:55:34.361619 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" event={"ID":"251ef657-822b-4392-9c92-266ccd5ca952","Type":"ContainerDied","Data":"762d7ae422f90cd7ab8b2856d50541644f647d79c9c4d20957c45e0fbf9075ef"} Dec 03 17:55:35 crc kubenswrapper[4787]: I1203 17:55:35.844176 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:55:35 crc kubenswrapper[4787]: I1203 17:55:35.992727 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-secret-0\") pod \"251ef657-822b-4392-9c92-266ccd5ca952\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " Dec 03 17:55:35 crc kubenswrapper[4787]: I1203 17:55:35.992872 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h6btg\" (UniqueName: \"kubernetes.io/projected/251ef657-822b-4392-9c92-266ccd5ca952-kube-api-access-h6btg\") pod \"251ef657-822b-4392-9c92-266ccd5ca952\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " Dec 03 17:55:35 crc kubenswrapper[4787]: I1203 17:55:35.993197 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-combined-ca-bundle\") pod \"251ef657-822b-4392-9c92-266ccd5ca952\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " Dec 03 17:55:35 crc kubenswrapper[4787]: I1203 17:55:35.993258 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-inventory\") pod \"251ef657-822b-4392-9c92-266ccd5ca952\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " Dec 03 17:55:35 crc kubenswrapper[4787]: I1203 17:55:35.993294 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-ssh-key\") pod \"251ef657-822b-4392-9c92-266ccd5ca952\" (UID: \"251ef657-822b-4392-9c92-266ccd5ca952\") " Dec 03 17:55:35 crc kubenswrapper[4787]: I1203 17:55:35.999487 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "251ef657-822b-4392-9c92-266ccd5ca952" (UID: "251ef657-822b-4392-9c92-266ccd5ca952"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:55:35 crc kubenswrapper[4787]: I1203 17:55:35.999490 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/251ef657-822b-4392-9c92-266ccd5ca952-kube-api-access-h6btg" (OuterVolumeSpecName: "kube-api-access-h6btg") pod "251ef657-822b-4392-9c92-266ccd5ca952" (UID: "251ef657-822b-4392-9c92-266ccd5ca952"). InnerVolumeSpecName "kube-api-access-h6btg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.020752 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-inventory" (OuterVolumeSpecName: "inventory") pod "251ef657-822b-4392-9c92-266ccd5ca952" (UID: "251ef657-822b-4392-9c92-266ccd5ca952"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.030146 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "251ef657-822b-4392-9c92-266ccd5ca952" (UID: "251ef657-822b-4392-9c92-266ccd5ca952"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.040139 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "251ef657-822b-4392-9c92-266ccd5ca952" (UID: "251ef657-822b-4392-9c92-266ccd5ca952"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.096767 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h6btg\" (UniqueName: \"kubernetes.io/projected/251ef657-822b-4392-9c92-266ccd5ca952-kube-api-access-h6btg\") on node \"crc\" DevicePath \"\"" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.096820 4787 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.096840 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.096858 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.096877 4787 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/251ef657-822b-4392-9c92-266ccd5ca952-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.394467 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" event={"ID":"251ef657-822b-4392-9c92-266ccd5ca952","Type":"ContainerDied","Data":"ee7051c0e9ed4c805082e13a6b6aa165879955e750f78578fdd3af8612940059"} Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.394502 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ee7051c0e9ed4c805082e13a6b6aa165879955e750f78578fdd3af8612940059" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.395130 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.534796 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk"] Dec 03 17:55:36 crc kubenswrapper[4787]: E1203 17:55:36.536173 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="251ef657-822b-4392-9c92-266ccd5ca952" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.536196 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="251ef657-822b-4392-9c92-266ccd5ca952" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.545883 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="251ef657-822b-4392-9c92-266ccd5ca952" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.547118 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.552146 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.554897 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk"] Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.554910 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.555323 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.555336 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.556741 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.718216 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.718313 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.718727 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.718971 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.719045 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.719174 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.719252 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbthb\" (UniqueName: \"kubernetes.io/projected/5ce09f0a-e6a4-4b04-a686-d43e1882f358-kube-api-access-lbthb\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.822013 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.822126 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.822190 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.822284 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbthb\" (UniqueName: \"kubernetes.io/projected/5ce09f0a-e6a4-4b04-a686-d43e1882f358-kube-api-access-lbthb\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.822466 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.822607 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.822768 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.826865 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.827476 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.827543 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.827698 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.828644 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.837147 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.844076 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbthb\" (UniqueName: \"kubernetes.io/projected/5ce09f0a-e6a4-4b04-a686-d43e1882f358-kube-api-access-lbthb\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:36 crc kubenswrapper[4787]: I1203 17:55:36.871996 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:55:37 crc kubenswrapper[4787]: I1203 17:55:37.423974 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk"] Dec 03 17:55:37 crc kubenswrapper[4787]: W1203 17:55:37.428175 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ce09f0a_e6a4_4b04_a686_d43e1882f358.slice/crio-17b38661f1ec8cb487f3662cf2e48bb378a03efd1ec2da4416e8606903c4405f WatchSource:0}: Error finding container 17b38661f1ec8cb487f3662cf2e48bb378a03efd1ec2da4416e8606903c4405f: Status 404 returned error can't find the container with id 17b38661f1ec8cb487f3662cf2e48bb378a03efd1ec2da4416e8606903c4405f Dec 03 17:55:37 crc kubenswrapper[4787]: I1203 17:55:37.432543 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 17:55:38 crc kubenswrapper[4787]: I1203 17:55:38.418499 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" event={"ID":"5ce09f0a-e6a4-4b04-a686-d43e1882f358","Type":"ContainerStarted","Data":"a82102220a8491ba74af32d29889eaab4a9504dec92e3b3f7feec792ca5b8c0e"} Dec 03 17:55:38 crc kubenswrapper[4787]: I1203 17:55:38.419074 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" event={"ID":"5ce09f0a-e6a4-4b04-a686-d43e1882f358","Type":"ContainerStarted","Data":"17b38661f1ec8cb487f3662cf2e48bb378a03efd1ec2da4416e8606903c4405f"} Dec 03 17:55:38 crc kubenswrapper[4787]: I1203 17:55:38.446685 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" podStartSLOduration=2.026638969 podStartE2EDuration="2.446667579s" podCreationTimestamp="2025-12-03 17:55:36 +0000 UTC" firstStartedPulling="2025-12-03 17:55:37.432342465 +0000 UTC m=+2554.249813424" lastFinishedPulling="2025-12-03 17:55:37.852371065 +0000 UTC m=+2554.669842034" observedRunningTime="2025-12-03 17:55:38.44185155 +0000 UTC m=+2555.259322589" watchObservedRunningTime="2025-12-03 17:55:38.446667579 +0000 UTC m=+2555.264138538" Dec 03 17:55:40 crc kubenswrapper[4787]: I1203 17:55:40.766550 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:55:40 crc kubenswrapper[4787]: E1203 17:55:40.767325 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:55:52 crc kubenswrapper[4787]: I1203 17:55:52.767391 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:55:52 crc kubenswrapper[4787]: E1203 17:55:52.768482 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:56:07 crc kubenswrapper[4787]: I1203 17:56:07.766118 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:56:07 crc kubenswrapper[4787]: E1203 17:56:07.766885 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:56:19 crc kubenswrapper[4787]: I1203 17:56:19.765795 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:56:19 crc kubenswrapper[4787]: E1203 17:56:19.767138 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:56:30 crc kubenswrapper[4787]: I1203 17:56:30.766586 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:56:30 crc kubenswrapper[4787]: E1203 17:56:30.767601 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:56:43 crc kubenswrapper[4787]: I1203 17:56:43.775537 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:56:43 crc kubenswrapper[4787]: E1203 17:56:43.777537 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:56:58 crc kubenswrapper[4787]: I1203 17:56:58.766386 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:56:58 crc kubenswrapper[4787]: E1203 17:56:58.767405 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:57:12 crc kubenswrapper[4787]: I1203 17:57:12.766595 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:57:12 crc kubenswrapper[4787]: E1203 17:57:12.767538 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:57:25 crc kubenswrapper[4787]: I1203 17:57:25.767109 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:57:25 crc kubenswrapper[4787]: E1203 17:57:25.768204 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:57:38 crc kubenswrapper[4787]: I1203 17:57:38.766418 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:57:38 crc kubenswrapper[4787]: E1203 17:57:38.767370 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:57:49 crc kubenswrapper[4787]: I1203 17:57:49.766218 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:57:49 crc kubenswrapper[4787]: E1203 17:57:49.767101 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:58:00 crc kubenswrapper[4787]: I1203 17:58:00.766399 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:58:00 crc kubenswrapper[4787]: E1203 17:58:00.767749 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.529702 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-k7rt5"] Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.533878 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.545638 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k7rt5"] Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.674651 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w9dj7\" (UniqueName: \"kubernetes.io/projected/82f70967-dc2d-4891-b183-691e8867b3bf-kube-api-access-w9dj7\") pod \"redhat-operators-k7rt5\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.674927 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-catalog-content\") pod \"redhat-operators-k7rt5\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.675183 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-utilities\") pod \"redhat-operators-k7rt5\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.777771 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w9dj7\" (UniqueName: \"kubernetes.io/projected/82f70967-dc2d-4891-b183-691e8867b3bf-kube-api-access-w9dj7\") pod \"redhat-operators-k7rt5\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.777836 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-catalog-content\") pod \"redhat-operators-k7rt5\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.778049 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-utilities\") pod \"redhat-operators-k7rt5\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.778725 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-utilities\") pod \"redhat-operators-k7rt5\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.779054 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-catalog-content\") pod \"redhat-operators-k7rt5\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.799914 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w9dj7\" (UniqueName: \"kubernetes.io/projected/82f70967-dc2d-4891-b183-691e8867b3bf-kube-api-access-w9dj7\") pod \"redhat-operators-k7rt5\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:02 crc kubenswrapper[4787]: I1203 17:58:02.855941 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:03 crc kubenswrapper[4787]: I1203 17:58:03.372088 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-k7rt5"] Dec 03 17:58:04 crc kubenswrapper[4787]: I1203 17:58:04.262977 4787 generic.go:334] "Generic (PLEG): container finished" podID="82f70967-dc2d-4891-b183-691e8867b3bf" containerID="f6f64a1ce499edd6f02f21219472a678ac207caf74400fd9d11a829c8b1a2cec" exitCode=0 Dec 03 17:58:04 crc kubenswrapper[4787]: I1203 17:58:04.263065 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7rt5" event={"ID":"82f70967-dc2d-4891-b183-691e8867b3bf","Type":"ContainerDied","Data":"f6f64a1ce499edd6f02f21219472a678ac207caf74400fd9d11a829c8b1a2cec"} Dec 03 17:58:04 crc kubenswrapper[4787]: I1203 17:58:04.263561 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7rt5" event={"ID":"82f70967-dc2d-4891-b183-691e8867b3bf","Type":"ContainerStarted","Data":"e8da5fa68ff0a57432ca1a50b5d8db2143caa82a7a4bde9e7118d5d0b60e143b"} Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.115919 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wgvvp"] Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.119104 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.138428 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-utilities\") pod \"certified-operators-wgvvp\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.138557 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-catalog-content\") pod \"certified-operators-wgvvp\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.138602 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psbb2\" (UniqueName: \"kubernetes.io/projected/9e8c9629-a77b-4532-ae02-73c84d4cf860-kube-api-access-psbb2\") pod \"certified-operators-wgvvp\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.159118 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wgvvp"] Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.240525 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-utilities\") pod \"certified-operators-wgvvp\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.240606 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-catalog-content\") pod \"certified-operators-wgvvp\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.240632 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psbb2\" (UniqueName: \"kubernetes.io/projected/9e8c9629-a77b-4532-ae02-73c84d4cf860-kube-api-access-psbb2\") pod \"certified-operators-wgvvp\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.241436 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-utilities\") pod \"certified-operators-wgvvp\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.241645 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-catalog-content\") pod \"certified-operators-wgvvp\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.264758 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psbb2\" (UniqueName: \"kubernetes.io/projected/9e8c9629-a77b-4532-ae02-73c84d4cf860-kube-api-access-psbb2\") pod \"certified-operators-wgvvp\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.498496 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:05 crc kubenswrapper[4787]: I1203 17:58:05.978713 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wgvvp"] Dec 03 17:58:06 crc kubenswrapper[4787]: I1203 17:58:06.288397 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7rt5" event={"ID":"82f70967-dc2d-4891-b183-691e8867b3bf","Type":"ContainerStarted","Data":"c9db5a8e20b23ce3b80b2d633a4342936f857debea91db5da0d252d2b4e09ba3"} Dec 03 17:58:06 crc kubenswrapper[4787]: I1203 17:58:06.290959 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgvvp" event={"ID":"9e8c9629-a77b-4532-ae02-73c84d4cf860","Type":"ContainerStarted","Data":"cee8f4e43f3f022babb217afb4ff313700865414a708337c7226a004f8a45636"} Dec 03 17:58:06 crc kubenswrapper[4787]: I1203 17:58:06.291002 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgvvp" event={"ID":"9e8c9629-a77b-4532-ae02-73c84d4cf860","Type":"ContainerStarted","Data":"abe7fdc4ecc633d484a3320c544bd1e1b18901bca5e1f72966eb6567e0179779"} Dec 03 17:58:07 crc kubenswrapper[4787]: I1203 17:58:07.303108 4787 generic.go:334] "Generic (PLEG): container finished" podID="9e8c9629-a77b-4532-ae02-73c84d4cf860" containerID="cee8f4e43f3f022babb217afb4ff313700865414a708337c7226a004f8a45636" exitCode=0 Dec 03 17:58:07 crc kubenswrapper[4787]: I1203 17:58:07.304288 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgvvp" event={"ID":"9e8c9629-a77b-4532-ae02-73c84d4cf860","Type":"ContainerDied","Data":"cee8f4e43f3f022babb217afb4ff313700865414a708337c7226a004f8a45636"} Dec 03 17:58:08 crc kubenswrapper[4787]: I1203 17:58:08.325213 4787 generic.go:334] "Generic (PLEG): container finished" podID="82f70967-dc2d-4891-b183-691e8867b3bf" containerID="c9db5a8e20b23ce3b80b2d633a4342936f857debea91db5da0d252d2b4e09ba3" exitCode=0 Dec 03 17:58:08 crc kubenswrapper[4787]: I1203 17:58:08.325307 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7rt5" event={"ID":"82f70967-dc2d-4891-b183-691e8867b3bf","Type":"ContainerDied","Data":"c9db5a8e20b23ce3b80b2d633a4342936f857debea91db5da0d252d2b4e09ba3"} Dec 03 17:58:09 crc kubenswrapper[4787]: I1203 17:58:09.341457 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7rt5" event={"ID":"82f70967-dc2d-4891-b183-691e8867b3bf","Type":"ContainerStarted","Data":"67264782f075ca1df3ec6acea31b68aac2054beb56f7ba46fa402b842086e695"} Dec 03 17:58:09 crc kubenswrapper[4787]: I1203 17:58:09.343619 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgvvp" event={"ID":"9e8c9629-a77b-4532-ae02-73c84d4cf860","Type":"ContainerStarted","Data":"27c35b4cda1a759c1dc5e0b6d58da536f6aa996655945fe5410a67c6ffa82e85"} Dec 03 17:58:09 crc kubenswrapper[4787]: I1203 17:58:09.376001 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-k7rt5" podStartSLOduration=2.7600661520000003 podStartE2EDuration="7.375983524s" podCreationTimestamp="2025-12-03 17:58:02 +0000 UTC" firstStartedPulling="2025-12-03 17:58:04.264880771 +0000 UTC m=+2701.082351730" lastFinishedPulling="2025-12-03 17:58:08.880798103 +0000 UTC m=+2705.698269102" observedRunningTime="2025-12-03 17:58:09.371142134 +0000 UTC m=+2706.188613103" watchObservedRunningTime="2025-12-03 17:58:09.375983524 +0000 UTC m=+2706.193454483" Dec 03 17:58:11 crc kubenswrapper[4787]: I1203 17:58:11.368421 4787 generic.go:334] "Generic (PLEG): container finished" podID="9e8c9629-a77b-4532-ae02-73c84d4cf860" containerID="27c35b4cda1a759c1dc5e0b6d58da536f6aa996655945fe5410a67c6ffa82e85" exitCode=0 Dec 03 17:58:11 crc kubenswrapper[4787]: I1203 17:58:11.368487 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgvvp" event={"ID":"9e8c9629-a77b-4532-ae02-73c84d4cf860","Type":"ContainerDied","Data":"27c35b4cda1a759c1dc5e0b6d58da536f6aa996655945fe5410a67c6ffa82e85"} Dec 03 17:58:12 crc kubenswrapper[4787]: I1203 17:58:12.382725 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgvvp" event={"ID":"9e8c9629-a77b-4532-ae02-73c84d4cf860","Type":"ContainerStarted","Data":"911c2c231a21d45fdcf046f5a38e5f79077707f70b25e56ebfdc3ced49799ec9"} Dec 03 17:58:12 crc kubenswrapper[4787]: I1203 17:58:12.409056 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wgvvp" podStartSLOduration=2.893569948 podStartE2EDuration="7.40903661s" podCreationTimestamp="2025-12-03 17:58:05 +0000 UTC" firstStartedPulling="2025-12-03 17:58:07.305964253 +0000 UTC m=+2704.123435232" lastFinishedPulling="2025-12-03 17:58:11.821430935 +0000 UTC m=+2708.638901894" observedRunningTime="2025-12-03 17:58:12.40308674 +0000 UTC m=+2709.220557709" watchObservedRunningTime="2025-12-03 17:58:12.40903661 +0000 UTC m=+2709.226507569" Dec 03 17:58:12 crc kubenswrapper[4787]: I1203 17:58:12.766104 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:58:12 crc kubenswrapper[4787]: E1203 17:58:12.766496 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 17:58:12 crc kubenswrapper[4787]: I1203 17:58:12.856160 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:12 crc kubenswrapper[4787]: I1203 17:58:12.856210 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:13 crc kubenswrapper[4787]: I1203 17:58:13.909505 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-k7rt5" podUID="82f70967-dc2d-4891-b183-691e8867b3bf" containerName="registry-server" probeResult="failure" output=< Dec 03 17:58:13 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 17:58:13 crc kubenswrapper[4787]: > Dec 03 17:58:15 crc kubenswrapper[4787]: I1203 17:58:15.502210 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:15 crc kubenswrapper[4787]: I1203 17:58:15.502529 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:15 crc kubenswrapper[4787]: I1203 17:58:15.566773 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:20 crc kubenswrapper[4787]: I1203 17:58:20.485501 4787 generic.go:334] "Generic (PLEG): container finished" podID="5ce09f0a-e6a4-4b04-a686-d43e1882f358" containerID="a82102220a8491ba74af32d29889eaab4a9504dec92e3b3f7feec792ca5b8c0e" exitCode=0 Dec 03 17:58:20 crc kubenswrapper[4787]: I1203 17:58:20.485587 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" event={"ID":"5ce09f0a-e6a4-4b04-a686-d43e1882f358","Type":"ContainerDied","Data":"a82102220a8491ba74af32d29889eaab4a9504dec92e3b3f7feec792ca5b8c0e"} Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.006950 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.135604 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-inventory\") pod \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.135675 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-telemetry-combined-ca-bundle\") pod \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.135706 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbthb\" (UniqueName: \"kubernetes.io/projected/5ce09f0a-e6a4-4b04-a686-d43e1882f358-kube-api-access-lbthb\") pod \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.135795 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-0\") pod \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.135827 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-1\") pod \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.135900 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ssh-key\") pod \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.135944 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-2\") pod \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\" (UID: \"5ce09f0a-e6a4-4b04-a686-d43e1882f358\") " Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.140874 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "5ce09f0a-e6a4-4b04-a686-d43e1882f358" (UID: "5ce09f0a-e6a4-4b04-a686-d43e1882f358"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.147385 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ce09f0a-e6a4-4b04-a686-d43e1882f358-kube-api-access-lbthb" (OuterVolumeSpecName: "kube-api-access-lbthb") pod "5ce09f0a-e6a4-4b04-a686-d43e1882f358" (UID: "5ce09f0a-e6a4-4b04-a686-d43e1882f358"). InnerVolumeSpecName "kube-api-access-lbthb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.170888 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "5ce09f0a-e6a4-4b04-a686-d43e1882f358" (UID: "5ce09f0a-e6a4-4b04-a686-d43e1882f358"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.171340 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5ce09f0a-e6a4-4b04-a686-d43e1882f358" (UID: "5ce09f0a-e6a4-4b04-a686-d43e1882f358"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.179158 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "5ce09f0a-e6a4-4b04-a686-d43e1882f358" (UID: "5ce09f0a-e6a4-4b04-a686-d43e1882f358"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.189862 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "5ce09f0a-e6a4-4b04-a686-d43e1882f358" (UID: "5ce09f0a-e6a4-4b04-a686-d43e1882f358"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.198857 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-inventory" (OuterVolumeSpecName: "inventory") pod "5ce09f0a-e6a4-4b04-a686-d43e1882f358" (UID: "5ce09f0a-e6a4-4b04-a686-d43e1882f358"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.238641 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.238680 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.238691 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.238701 4787 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.238712 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbthb\" (UniqueName: \"kubernetes.io/projected/5ce09f0a-e6a4-4b04-a686-d43e1882f358-kube-api-access-lbthb\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.238721 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.238729 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/5ce09f0a-e6a4-4b04-a686-d43e1882f358-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.509970 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" event={"ID":"5ce09f0a-e6a4-4b04-a686-d43e1882f358","Type":"ContainerDied","Data":"17b38661f1ec8cb487f3662cf2e48bb378a03efd1ec2da4416e8606903c4405f"} Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.510039 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17b38661f1ec8cb487f3662cf2e48bb378a03efd1ec2da4416e8606903c4405f" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.510089 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.661481 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf"] Dec 03 17:58:22 crc kubenswrapper[4787]: E1203 17:58:22.662237 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ce09f0a-e6a4-4b04-a686-d43e1882f358" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.662261 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ce09f0a-e6a4-4b04-a686-d43e1882f358" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.662652 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ce09f0a-e6a4-4b04-a686-d43e1882f358" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.663639 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.667841 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.667936 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.667949 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.668258 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.668309 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.694395 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf"] Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.753101 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j7442\" (UniqueName: \"kubernetes.io/projected/dfdfd706-a8f9-49ab-aba5-514d80a90c96-kube-api-access-j7442\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.753164 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.753407 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.753625 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.753895 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.754141 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.754225 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.855922 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j7442\" (UniqueName: \"kubernetes.io/projected/dfdfd706-a8f9-49ab-aba5-514d80a90c96-kube-api-access-j7442\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.855985 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.856063 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.856122 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.856224 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.856316 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.856368 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.863091 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.863194 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.863943 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.864500 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.864640 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.866220 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.874545 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j7442\" (UniqueName: \"kubernetes.io/projected/dfdfd706-a8f9-49ab-aba5-514d80a90c96-kube-api-access-j7442\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.917252 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.986439 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 17:58:22 crc kubenswrapper[4787]: I1203 17:58:22.987614 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:23 crc kubenswrapper[4787]: I1203 17:58:23.174294 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k7rt5"] Dec 03 17:58:23 crc kubenswrapper[4787]: I1203 17:58:23.536710 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf"] Dec 03 17:58:23 crc kubenswrapper[4787]: W1203 17:58:23.541887 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfdfd706_a8f9_49ab_aba5_514d80a90c96.slice/crio-f7283604c64a009d71276a4521622252e09ac1d94a91826a07d014dd818b62b5 WatchSource:0}: Error finding container f7283604c64a009d71276a4521622252e09ac1d94a91826a07d014dd818b62b5: Status 404 returned error can't find the container with id f7283604c64a009d71276a4521622252e09ac1d94a91826a07d014dd818b62b5 Dec 03 17:58:24 crc kubenswrapper[4787]: I1203 17:58:24.529813 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" event={"ID":"dfdfd706-a8f9-49ab-aba5-514d80a90c96","Type":"ContainerStarted","Data":"f7283604c64a009d71276a4521622252e09ac1d94a91826a07d014dd818b62b5"} Dec 03 17:58:24 crc kubenswrapper[4787]: I1203 17:58:24.530040 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-k7rt5" podUID="82f70967-dc2d-4891-b183-691e8867b3bf" containerName="registry-server" containerID="cri-o://67264782f075ca1df3ec6acea31b68aac2054beb56f7ba46fa402b842086e695" gracePeriod=2 Dec 03 17:58:25 crc kubenswrapper[4787]: I1203 17:58:25.547879 4787 generic.go:334] "Generic (PLEG): container finished" podID="82f70967-dc2d-4891-b183-691e8867b3bf" containerID="67264782f075ca1df3ec6acea31b68aac2054beb56f7ba46fa402b842086e695" exitCode=0 Dec 03 17:58:25 crc kubenswrapper[4787]: I1203 17:58:25.548238 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7rt5" event={"ID":"82f70967-dc2d-4891-b183-691e8867b3bf","Type":"ContainerDied","Data":"67264782f075ca1df3ec6acea31b68aac2054beb56f7ba46fa402b842086e695"} Dec 03 17:58:25 crc kubenswrapper[4787]: I1203 17:58:25.552850 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" event={"ID":"dfdfd706-a8f9-49ab-aba5-514d80a90c96","Type":"ContainerStarted","Data":"52bac3d60506d1bb40b46024f3fbe02eb06329e1ec019f55628e2397d5d3062e"} Dec 03 17:58:25 crc kubenswrapper[4787]: I1203 17:58:25.556223 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:25 crc kubenswrapper[4787]: I1203 17:58:25.583131 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" podStartSLOduration=2.719987347 podStartE2EDuration="3.583110027s" podCreationTimestamp="2025-12-03 17:58:22 +0000 UTC" firstStartedPulling="2025-12-03 17:58:23.543904706 +0000 UTC m=+2720.361375665" lastFinishedPulling="2025-12-03 17:58:24.407027396 +0000 UTC m=+2721.224498345" observedRunningTime="2025-12-03 17:58:25.577666421 +0000 UTC m=+2722.395137420" watchObservedRunningTime="2025-12-03 17:58:25.583110027 +0000 UTC m=+2722.400580996" Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.129153 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.232495 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-utilities\") pod \"82f70967-dc2d-4891-b183-691e8867b3bf\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.233123 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9dj7\" (UniqueName: \"kubernetes.io/projected/82f70967-dc2d-4891-b183-691e8867b3bf-kube-api-access-w9dj7\") pod \"82f70967-dc2d-4891-b183-691e8867b3bf\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.233134 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-utilities" (OuterVolumeSpecName: "utilities") pod "82f70967-dc2d-4891-b183-691e8867b3bf" (UID: "82f70967-dc2d-4891-b183-691e8867b3bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.233330 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-catalog-content\") pod \"82f70967-dc2d-4891-b183-691e8867b3bf\" (UID: \"82f70967-dc2d-4891-b183-691e8867b3bf\") " Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.234372 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.244645 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/82f70967-dc2d-4891-b183-691e8867b3bf-kube-api-access-w9dj7" (OuterVolumeSpecName: "kube-api-access-w9dj7") pod "82f70967-dc2d-4891-b183-691e8867b3bf" (UID: "82f70967-dc2d-4891-b183-691e8867b3bf"). InnerVolumeSpecName "kube-api-access-w9dj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.337693 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9dj7\" (UniqueName: \"kubernetes.io/projected/82f70967-dc2d-4891-b183-691e8867b3bf-kube-api-access-w9dj7\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.356180 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "82f70967-dc2d-4891-b183-691e8867b3bf" (UID: "82f70967-dc2d-4891-b183-691e8867b3bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.440125 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/82f70967-dc2d-4891-b183-691e8867b3bf-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.569328 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-k7rt5" event={"ID":"82f70967-dc2d-4891-b183-691e8867b3bf","Type":"ContainerDied","Data":"e8da5fa68ff0a57432ca1a50b5d8db2143caa82a7a4bde9e7118d5d0b60e143b"} Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.569376 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-k7rt5" Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.569409 4787 scope.go:117] "RemoveContainer" containerID="67264782f075ca1df3ec6acea31b68aac2054beb56f7ba46fa402b842086e695" Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.606426 4787 scope.go:117] "RemoveContainer" containerID="c9db5a8e20b23ce3b80b2d633a4342936f857debea91db5da0d252d2b4e09ba3" Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.611833 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-k7rt5"] Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.623903 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-k7rt5"] Dec 03 17:58:26 crc kubenswrapper[4787]: I1203 17:58:26.646428 4787 scope.go:117] "RemoveContainer" containerID="f6f64a1ce499edd6f02f21219472a678ac207caf74400fd9d11a829c8b1a2cec" Dec 03 17:58:27 crc kubenswrapper[4787]: I1203 17:58:27.363215 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wgvvp"] Dec 03 17:58:27 crc kubenswrapper[4787]: I1203 17:58:27.363779 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-wgvvp" podUID="9e8c9629-a77b-4532-ae02-73c84d4cf860" containerName="registry-server" containerID="cri-o://911c2c231a21d45fdcf046f5a38e5f79077707f70b25e56ebfdc3ced49799ec9" gracePeriod=2 Dec 03 17:58:27 crc kubenswrapper[4787]: I1203 17:58:27.583968 4787 generic.go:334] "Generic (PLEG): container finished" podID="9e8c9629-a77b-4532-ae02-73c84d4cf860" containerID="911c2c231a21d45fdcf046f5a38e5f79077707f70b25e56ebfdc3ced49799ec9" exitCode=0 Dec 03 17:58:27 crc kubenswrapper[4787]: I1203 17:58:27.584007 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgvvp" event={"ID":"9e8c9629-a77b-4532-ae02-73c84d4cf860","Type":"ContainerDied","Data":"911c2c231a21d45fdcf046f5a38e5f79077707f70b25e56ebfdc3ced49799ec9"} Dec 03 17:58:27 crc kubenswrapper[4787]: I1203 17:58:27.766412 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 17:58:27 crc kubenswrapper[4787]: I1203 17:58:27.779128 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="82f70967-dc2d-4891-b183-691e8867b3bf" path="/var/lib/kubelet/pods/82f70967-dc2d-4891-b183-691e8867b3bf/volumes" Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.387392 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.481501 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-catalog-content\") pod \"9e8c9629-a77b-4532-ae02-73c84d4cf860\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.481685 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-utilities\") pod \"9e8c9629-a77b-4532-ae02-73c84d4cf860\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.481780 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-psbb2\" (UniqueName: \"kubernetes.io/projected/9e8c9629-a77b-4532-ae02-73c84d4cf860-kube-api-access-psbb2\") pod \"9e8c9629-a77b-4532-ae02-73c84d4cf860\" (UID: \"9e8c9629-a77b-4532-ae02-73c84d4cf860\") " Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.482798 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-utilities" (OuterVolumeSpecName: "utilities") pod "9e8c9629-a77b-4532-ae02-73c84d4cf860" (UID: "9e8c9629-a77b-4532-ae02-73c84d4cf860"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.493158 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e8c9629-a77b-4532-ae02-73c84d4cf860-kube-api-access-psbb2" (OuterVolumeSpecName: "kube-api-access-psbb2") pod "9e8c9629-a77b-4532-ae02-73c84d4cf860" (UID: "9e8c9629-a77b-4532-ae02-73c84d4cf860"). InnerVolumeSpecName "kube-api-access-psbb2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.536535 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9e8c9629-a77b-4532-ae02-73c84d4cf860" (UID: "9e8c9629-a77b-4532-ae02-73c84d4cf860"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.585349 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.585382 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9e8c9629-a77b-4532-ae02-73c84d4cf860-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.585395 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-psbb2\" (UniqueName: \"kubernetes.io/projected/9e8c9629-a77b-4532-ae02-73c84d4cf860-kube-api-access-psbb2\") on node \"crc\" DevicePath \"\"" Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.624041 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wgvvp" event={"ID":"9e8c9629-a77b-4532-ae02-73c84d4cf860","Type":"ContainerDied","Data":"abe7fdc4ecc633d484a3320c544bd1e1b18901bca5e1f72966eb6567e0179779"} Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.624094 4787 scope.go:117] "RemoveContainer" containerID="911c2c231a21d45fdcf046f5a38e5f79077707f70b25e56ebfdc3ced49799ec9" Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.624182 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wgvvp" Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.672290 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"5a433e35c83e9c38d50149fbfb3dd5416667750d83bba18e735da69f228bbfe5"} Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.762190 4787 scope.go:117] "RemoveContainer" containerID="27c35b4cda1a759c1dc5e0b6d58da536f6aa996655945fe5410a67c6ffa82e85" Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.778934 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-wgvvp"] Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.797893 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-wgvvp"] Dec 03 17:58:28 crc kubenswrapper[4787]: I1203 17:58:28.821546 4787 scope.go:117] "RemoveContainer" containerID="cee8f4e43f3f022babb217afb4ff313700865414a708337c7226a004f8a45636" Dec 03 17:58:29 crc kubenswrapper[4787]: I1203 17:58:29.786912 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e8c9629-a77b-4532-ae02-73c84d4cf860" path="/var/lib/kubelet/pods/9e8c9629-a77b-4532-ae02-73c84d4cf860/volumes" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.188697 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-2kvtl"] Dec 03 17:58:53 crc kubenswrapper[4787]: E1203 17:58:53.189781 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f70967-dc2d-4891-b183-691e8867b3bf" containerName="extract-utilities" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.189800 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f70967-dc2d-4891-b183-691e8867b3bf" containerName="extract-utilities" Dec 03 17:58:53 crc kubenswrapper[4787]: E1203 17:58:53.189818 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8c9629-a77b-4532-ae02-73c84d4cf860" containerName="extract-content" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.189827 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8c9629-a77b-4532-ae02-73c84d4cf860" containerName="extract-content" Dec 03 17:58:53 crc kubenswrapper[4787]: E1203 17:58:53.189840 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f70967-dc2d-4891-b183-691e8867b3bf" containerName="extract-content" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.189850 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f70967-dc2d-4891-b183-691e8867b3bf" containerName="extract-content" Dec 03 17:58:53 crc kubenswrapper[4787]: E1203 17:58:53.189866 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8c9629-a77b-4532-ae02-73c84d4cf860" containerName="registry-server" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.189874 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8c9629-a77b-4532-ae02-73c84d4cf860" containerName="registry-server" Dec 03 17:58:53 crc kubenswrapper[4787]: E1203 17:58:53.189904 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="82f70967-dc2d-4891-b183-691e8867b3bf" containerName="registry-server" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.189912 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="82f70967-dc2d-4891-b183-691e8867b3bf" containerName="registry-server" Dec 03 17:58:53 crc kubenswrapper[4787]: E1203 17:58:53.189934 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e8c9629-a77b-4532-ae02-73c84d4cf860" containerName="extract-utilities" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.189942 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e8c9629-a77b-4532-ae02-73c84d4cf860" containerName="extract-utilities" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.190209 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="82f70967-dc2d-4891-b183-691e8867b3bf" containerName="registry-server" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.190239 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e8c9629-a77b-4532-ae02-73c84d4cf860" containerName="registry-server" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.193262 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.209847 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kvtl"] Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.258081 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-utilities\") pod \"redhat-marketplace-2kvtl\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.258370 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k5xbt\" (UniqueName: \"kubernetes.io/projected/636371b6-4448-4961-83f7-cf4862dc5c54-kube-api-access-k5xbt\") pod \"redhat-marketplace-2kvtl\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.258556 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-catalog-content\") pod \"redhat-marketplace-2kvtl\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.360149 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-utilities\") pod \"redhat-marketplace-2kvtl\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.360660 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-utilities\") pod \"redhat-marketplace-2kvtl\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.360726 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k5xbt\" (UniqueName: \"kubernetes.io/projected/636371b6-4448-4961-83f7-cf4862dc5c54-kube-api-access-k5xbt\") pod \"redhat-marketplace-2kvtl\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.360889 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-catalog-content\") pod \"redhat-marketplace-2kvtl\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.361154 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-catalog-content\") pod \"redhat-marketplace-2kvtl\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.384987 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k5xbt\" (UniqueName: \"kubernetes.io/projected/636371b6-4448-4961-83f7-cf4862dc5c54-kube-api-access-k5xbt\") pod \"redhat-marketplace-2kvtl\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:58:53 crc kubenswrapper[4787]: I1203 17:58:53.511201 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:58:54 crc kubenswrapper[4787]: I1203 17:58:54.002534 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kvtl"] Dec 03 17:58:54 crc kubenswrapper[4787]: W1203 17:58:54.012623 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod636371b6_4448_4961_83f7_cf4862dc5c54.slice/crio-c49b35fe6511a1fc96324c738ae53d02dbd63bd80918667f351deeb649b3608a WatchSource:0}: Error finding container c49b35fe6511a1fc96324c738ae53d02dbd63bd80918667f351deeb649b3608a: Status 404 returned error can't find the container with id c49b35fe6511a1fc96324c738ae53d02dbd63bd80918667f351deeb649b3608a Dec 03 17:58:54 crc kubenswrapper[4787]: I1203 17:58:54.963358 4787 generic.go:334] "Generic (PLEG): container finished" podID="636371b6-4448-4961-83f7-cf4862dc5c54" containerID="8dbfbacaeb89f4d65af7c808852b2e6e19f164746bf0c640eae782ebc8184046" exitCode=0 Dec 03 17:58:54 crc kubenswrapper[4787]: I1203 17:58:54.963773 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kvtl" event={"ID":"636371b6-4448-4961-83f7-cf4862dc5c54","Type":"ContainerDied","Data":"8dbfbacaeb89f4d65af7c808852b2e6e19f164746bf0c640eae782ebc8184046"} Dec 03 17:58:54 crc kubenswrapper[4787]: I1203 17:58:54.964945 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kvtl" event={"ID":"636371b6-4448-4961-83f7-cf4862dc5c54","Type":"ContainerStarted","Data":"c49b35fe6511a1fc96324c738ae53d02dbd63bd80918667f351deeb649b3608a"} Dec 03 17:58:56 crc kubenswrapper[4787]: I1203 17:58:56.986949 4787 generic.go:334] "Generic (PLEG): container finished" podID="636371b6-4448-4961-83f7-cf4862dc5c54" containerID="26bf13a37ead071fad16f66fd4a2f7d5ebefd064762548671bf9e12d76768b02" exitCode=0 Dec 03 17:58:56 crc kubenswrapper[4787]: I1203 17:58:56.987041 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kvtl" event={"ID":"636371b6-4448-4961-83f7-cf4862dc5c54","Type":"ContainerDied","Data":"26bf13a37ead071fad16f66fd4a2f7d5ebefd064762548671bf9e12d76768b02"} Dec 03 17:58:59 crc kubenswrapper[4787]: I1203 17:58:59.021117 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kvtl" event={"ID":"636371b6-4448-4961-83f7-cf4862dc5c54","Type":"ContainerStarted","Data":"17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320"} Dec 03 17:58:59 crc kubenswrapper[4787]: I1203 17:58:59.052996 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-2kvtl" podStartSLOduration=3.204693937 podStartE2EDuration="6.052970472s" podCreationTimestamp="2025-12-03 17:58:53 +0000 UTC" firstStartedPulling="2025-12-03 17:58:54.966869915 +0000 UTC m=+2751.784340864" lastFinishedPulling="2025-12-03 17:58:57.81514644 +0000 UTC m=+2754.632617399" observedRunningTime="2025-12-03 17:58:59.04206279 +0000 UTC m=+2755.859533819" watchObservedRunningTime="2025-12-03 17:58:59.052970472 +0000 UTC m=+2755.870441461" Dec 03 17:59:03 crc kubenswrapper[4787]: I1203 17:59:03.512089 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:59:03 crc kubenswrapper[4787]: I1203 17:59:03.512675 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:59:03 crc kubenswrapper[4787]: I1203 17:59:03.573646 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:59:04 crc kubenswrapper[4787]: I1203 17:59:04.170469 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:59:04 crc kubenswrapper[4787]: I1203 17:59:04.243719 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kvtl"] Dec 03 17:59:06 crc kubenswrapper[4787]: I1203 17:59:06.101242 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-2kvtl" podUID="636371b6-4448-4961-83f7-cf4862dc5c54" containerName="registry-server" containerID="cri-o://17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320" gracePeriod=2 Dec 03 17:59:06 crc kubenswrapper[4787]: I1203 17:59:06.672516 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:59:06 crc kubenswrapper[4787]: I1203 17:59:06.862954 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-catalog-content\") pod \"636371b6-4448-4961-83f7-cf4862dc5c54\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " Dec 03 17:59:06 crc kubenswrapper[4787]: I1203 17:59:06.863301 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-utilities\") pod \"636371b6-4448-4961-83f7-cf4862dc5c54\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " Dec 03 17:59:06 crc kubenswrapper[4787]: I1203 17:59:06.863369 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k5xbt\" (UniqueName: \"kubernetes.io/projected/636371b6-4448-4961-83f7-cf4862dc5c54-kube-api-access-k5xbt\") pod \"636371b6-4448-4961-83f7-cf4862dc5c54\" (UID: \"636371b6-4448-4961-83f7-cf4862dc5c54\") " Dec 03 17:59:06 crc kubenswrapper[4787]: I1203 17:59:06.864775 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-utilities" (OuterVolumeSpecName: "utilities") pod "636371b6-4448-4961-83f7-cf4862dc5c54" (UID: "636371b6-4448-4961-83f7-cf4862dc5c54"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:59:06 crc kubenswrapper[4787]: I1203 17:59:06.872269 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/636371b6-4448-4961-83f7-cf4862dc5c54-kube-api-access-k5xbt" (OuterVolumeSpecName: "kube-api-access-k5xbt") pod "636371b6-4448-4961-83f7-cf4862dc5c54" (UID: "636371b6-4448-4961-83f7-cf4862dc5c54"). InnerVolumeSpecName "kube-api-access-k5xbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 17:59:06 crc kubenswrapper[4787]: I1203 17:59:06.886164 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "636371b6-4448-4961-83f7-cf4862dc5c54" (UID: "636371b6-4448-4961-83f7-cf4862dc5c54"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 17:59:06 crc kubenswrapper[4787]: I1203 17:59:06.965991 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 17:59:06 crc kubenswrapper[4787]: I1203 17:59:06.966062 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k5xbt\" (UniqueName: \"kubernetes.io/projected/636371b6-4448-4961-83f7-cf4862dc5c54-kube-api-access-k5xbt\") on node \"crc\" DevicePath \"\"" Dec 03 17:59:06 crc kubenswrapper[4787]: I1203 17:59:06.966079 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/636371b6-4448-4961-83f7-cf4862dc5c54-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.115158 4787 generic.go:334] "Generic (PLEG): container finished" podID="636371b6-4448-4961-83f7-cf4862dc5c54" containerID="17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320" exitCode=0 Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.115226 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-2kvtl" Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.115233 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kvtl" event={"ID":"636371b6-4448-4961-83f7-cf4862dc5c54","Type":"ContainerDied","Data":"17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320"} Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.115326 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-2kvtl" event={"ID":"636371b6-4448-4961-83f7-cf4862dc5c54","Type":"ContainerDied","Data":"c49b35fe6511a1fc96324c738ae53d02dbd63bd80918667f351deeb649b3608a"} Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.115355 4787 scope.go:117] "RemoveContainer" containerID="17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320" Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.149684 4787 scope.go:117] "RemoveContainer" containerID="26bf13a37ead071fad16f66fd4a2f7d5ebefd064762548671bf9e12d76768b02" Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.156414 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kvtl"] Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.173644 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-2kvtl"] Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.179981 4787 scope.go:117] "RemoveContainer" containerID="8dbfbacaeb89f4d65af7c808852b2e6e19f164746bf0c640eae782ebc8184046" Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.264999 4787 scope.go:117] "RemoveContainer" containerID="17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320" Dec 03 17:59:07 crc kubenswrapper[4787]: E1203 17:59:07.265504 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320\": container with ID starting with 17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320 not found: ID does not exist" containerID="17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320" Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.265542 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320"} err="failed to get container status \"17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320\": rpc error: code = NotFound desc = could not find container \"17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320\": container with ID starting with 17b580e241db105d62612961a73a637ecf93f3379838bfce911ed394b4494320 not found: ID does not exist" Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.265566 4787 scope.go:117] "RemoveContainer" containerID="26bf13a37ead071fad16f66fd4a2f7d5ebefd064762548671bf9e12d76768b02" Dec 03 17:59:07 crc kubenswrapper[4787]: E1203 17:59:07.265859 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26bf13a37ead071fad16f66fd4a2f7d5ebefd064762548671bf9e12d76768b02\": container with ID starting with 26bf13a37ead071fad16f66fd4a2f7d5ebefd064762548671bf9e12d76768b02 not found: ID does not exist" containerID="26bf13a37ead071fad16f66fd4a2f7d5ebefd064762548671bf9e12d76768b02" Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.265898 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26bf13a37ead071fad16f66fd4a2f7d5ebefd064762548671bf9e12d76768b02"} err="failed to get container status \"26bf13a37ead071fad16f66fd4a2f7d5ebefd064762548671bf9e12d76768b02\": rpc error: code = NotFound desc = could not find container \"26bf13a37ead071fad16f66fd4a2f7d5ebefd064762548671bf9e12d76768b02\": container with ID starting with 26bf13a37ead071fad16f66fd4a2f7d5ebefd064762548671bf9e12d76768b02 not found: ID does not exist" Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.265965 4787 scope.go:117] "RemoveContainer" containerID="8dbfbacaeb89f4d65af7c808852b2e6e19f164746bf0c640eae782ebc8184046" Dec 03 17:59:07 crc kubenswrapper[4787]: E1203 17:59:07.266463 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8dbfbacaeb89f4d65af7c808852b2e6e19f164746bf0c640eae782ebc8184046\": container with ID starting with 8dbfbacaeb89f4d65af7c808852b2e6e19f164746bf0c640eae782ebc8184046 not found: ID does not exist" containerID="8dbfbacaeb89f4d65af7c808852b2e6e19f164746bf0c640eae782ebc8184046" Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.266495 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8dbfbacaeb89f4d65af7c808852b2e6e19f164746bf0c640eae782ebc8184046"} err="failed to get container status \"8dbfbacaeb89f4d65af7c808852b2e6e19f164746bf0c640eae782ebc8184046\": rpc error: code = NotFound desc = could not find container \"8dbfbacaeb89f4d65af7c808852b2e6e19f164746bf0c640eae782ebc8184046\": container with ID starting with 8dbfbacaeb89f4d65af7c808852b2e6e19f164746bf0c640eae782ebc8184046 not found: ID does not exist" Dec 03 17:59:07 crc kubenswrapper[4787]: I1203 17:59:07.778723 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="636371b6-4448-4961-83f7-cf4862dc5c54" path="/var/lib/kubelet/pods/636371b6-4448-4961-83f7-cf4862dc5c54/volumes" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.169676 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2"] Dec 03 18:00:00 crc kubenswrapper[4787]: E1203 18:00:00.171179 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636371b6-4448-4961-83f7-cf4862dc5c54" containerName="extract-utilities" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.171208 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="636371b6-4448-4961-83f7-cf4862dc5c54" containerName="extract-utilities" Dec 03 18:00:00 crc kubenswrapper[4787]: E1203 18:00:00.171259 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636371b6-4448-4961-83f7-cf4862dc5c54" containerName="registry-server" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.171274 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="636371b6-4448-4961-83f7-cf4862dc5c54" containerName="registry-server" Dec 03 18:00:00 crc kubenswrapper[4787]: E1203 18:00:00.171311 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="636371b6-4448-4961-83f7-cf4862dc5c54" containerName="extract-content" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.171324 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="636371b6-4448-4961-83f7-cf4862dc5c54" containerName="extract-content" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.171705 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="636371b6-4448-4961-83f7-cf4862dc5c54" containerName="registry-server" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.173326 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.176122 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.176179 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.180877 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2"] Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.301248 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f03a6f41-66de-4c21-84a1-8c30b66c80b8-config-volume\") pod \"collect-profiles-29413080-9hmf2\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.301353 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f03a6f41-66de-4c21-84a1-8c30b66c80b8-secret-volume\") pod \"collect-profiles-29413080-9hmf2\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.301380 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cl7w\" (UniqueName: \"kubernetes.io/projected/f03a6f41-66de-4c21-84a1-8c30b66c80b8-kube-api-access-8cl7w\") pod \"collect-profiles-29413080-9hmf2\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.404074 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f03a6f41-66de-4c21-84a1-8c30b66c80b8-secret-volume\") pod \"collect-profiles-29413080-9hmf2\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.404405 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cl7w\" (UniqueName: \"kubernetes.io/projected/f03a6f41-66de-4c21-84a1-8c30b66c80b8-kube-api-access-8cl7w\") pod \"collect-profiles-29413080-9hmf2\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.404653 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f03a6f41-66de-4c21-84a1-8c30b66c80b8-config-volume\") pod \"collect-profiles-29413080-9hmf2\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.405675 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f03a6f41-66de-4c21-84a1-8c30b66c80b8-config-volume\") pod \"collect-profiles-29413080-9hmf2\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.411249 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f03a6f41-66de-4c21-84a1-8c30b66c80b8-secret-volume\") pod \"collect-profiles-29413080-9hmf2\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.423051 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cl7w\" (UniqueName: \"kubernetes.io/projected/f03a6f41-66de-4c21-84a1-8c30b66c80b8-kube-api-access-8cl7w\") pod \"collect-profiles-29413080-9hmf2\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.495909 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:00 crc kubenswrapper[4787]: I1203 18:00:00.967716 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2"] Dec 03 18:00:01 crc kubenswrapper[4787]: I1203 18:00:01.729885 4787 generic.go:334] "Generic (PLEG): container finished" podID="f03a6f41-66de-4c21-84a1-8c30b66c80b8" containerID="14867a714d1ee66f92e7c58dd907908da0d9d440efc984f967c8903efd0addf3" exitCode=0 Dec 03 18:00:01 crc kubenswrapper[4787]: I1203 18:00:01.730152 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" event={"ID":"f03a6f41-66de-4c21-84a1-8c30b66c80b8","Type":"ContainerDied","Data":"14867a714d1ee66f92e7c58dd907908da0d9d440efc984f967c8903efd0addf3"} Dec 03 18:00:01 crc kubenswrapper[4787]: I1203 18:00:01.730183 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" event={"ID":"f03a6f41-66de-4c21-84a1-8c30b66c80b8","Type":"ContainerStarted","Data":"c27babdd468ed5ac6e4ee2b5e121475f105d11c580cc49879f629dabad408d2e"} Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.157773 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.266854 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cl7w\" (UniqueName: \"kubernetes.io/projected/f03a6f41-66de-4c21-84a1-8c30b66c80b8-kube-api-access-8cl7w\") pod \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.267209 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f03a6f41-66de-4c21-84a1-8c30b66c80b8-config-volume\") pod \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.267366 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f03a6f41-66de-4c21-84a1-8c30b66c80b8-secret-volume\") pod \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\" (UID: \"f03a6f41-66de-4c21-84a1-8c30b66c80b8\") " Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.267889 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f03a6f41-66de-4c21-84a1-8c30b66c80b8-config-volume" (OuterVolumeSpecName: "config-volume") pod "f03a6f41-66de-4c21-84a1-8c30b66c80b8" (UID: "f03a6f41-66de-4c21-84a1-8c30b66c80b8"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.268122 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f03a6f41-66de-4c21-84a1-8c30b66c80b8-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.275291 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f03a6f41-66de-4c21-84a1-8c30b66c80b8-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f03a6f41-66de-4c21-84a1-8c30b66c80b8" (UID: "f03a6f41-66de-4c21-84a1-8c30b66c80b8"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.275403 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f03a6f41-66de-4c21-84a1-8c30b66c80b8-kube-api-access-8cl7w" (OuterVolumeSpecName: "kube-api-access-8cl7w") pod "f03a6f41-66de-4c21-84a1-8c30b66c80b8" (UID: "f03a6f41-66de-4c21-84a1-8c30b66c80b8"). InnerVolumeSpecName "kube-api-access-8cl7w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.370628 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f03a6f41-66de-4c21-84a1-8c30b66c80b8-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.370667 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cl7w\" (UniqueName: \"kubernetes.io/projected/f03a6f41-66de-4c21-84a1-8c30b66c80b8-kube-api-access-8cl7w\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.755738 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" event={"ID":"f03a6f41-66de-4c21-84a1-8c30b66c80b8","Type":"ContainerDied","Data":"c27babdd468ed5ac6e4ee2b5e121475f105d11c580cc49879f629dabad408d2e"} Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.756122 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c27babdd468ed5ac6e4ee2b5e121475f105d11c580cc49879f629dabad408d2e" Dec 03 18:00:03 crc kubenswrapper[4787]: I1203 18:00:03.755804 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2" Dec 03 18:00:04 crc kubenswrapper[4787]: I1203 18:00:04.247263 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb"] Dec 03 18:00:04 crc kubenswrapper[4787]: I1203 18:00:04.267774 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413035-tghvb"] Dec 03 18:00:05 crc kubenswrapper[4787]: I1203 18:00:05.788247 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75" path="/var/lib/kubelet/pods/4627ba1f-acfb-4fbb-bf42-6b0dbbc0de75/volumes" Dec 03 18:00:07 crc kubenswrapper[4787]: E1203 18:00:07.042995 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice/crio-c27babdd468ed5ac6e4ee2b5e121475f105d11c580cc49879f629dabad408d2e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice\": RecentStats: unable to find data in memory cache]" Dec 03 18:00:17 crc kubenswrapper[4787]: E1203 18:00:17.405284 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice/crio-c27babdd468ed5ac6e4ee2b5e121475f105d11c580cc49879f629dabad408d2e\": RecentStats: unable to find data in memory cache]" Dec 03 18:00:25 crc kubenswrapper[4787]: I1203 18:00:25.285973 4787 scope.go:117] "RemoveContainer" containerID="bd518d37cbfe99df5fa35cdf85b249e1185deb89579c21058d3dbc263fedd71e" Dec 03 18:00:27 crc kubenswrapper[4787]: E1203 18:00:27.695288 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice/crio-c27babdd468ed5ac6e4ee2b5e121475f105d11c580cc49879f629dabad408d2e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice\": RecentStats: unable to find data in memory cache]" Dec 03 18:00:37 crc kubenswrapper[4787]: I1203 18:00:37.115668 4787 generic.go:334] "Generic (PLEG): container finished" podID="dfdfd706-a8f9-49ab-aba5-514d80a90c96" containerID="52bac3d60506d1bb40b46024f3fbe02eb06329e1ec019f55628e2397d5d3062e" exitCode=0 Dec 03 18:00:37 crc kubenswrapper[4787]: I1203 18:00:37.115792 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" event={"ID":"dfdfd706-a8f9-49ab-aba5-514d80a90c96","Type":"ContainerDied","Data":"52bac3d60506d1bb40b46024f3fbe02eb06329e1ec019f55628e2397d5d3062e"} Dec 03 18:00:37 crc kubenswrapper[4787]: E1203 18:00:37.994935 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice/crio-c27babdd468ed5ac6e4ee2b5e121475f105d11c580cc49879f629dabad408d2e\": RecentStats: unable to find data in memory cache]" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.548154 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.675832 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j7442\" (UniqueName: \"kubernetes.io/projected/dfdfd706-a8f9-49ab-aba5-514d80a90c96-kube-api-access-j7442\") pod \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.675929 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ssh-key\") pod \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.675996 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-1\") pod \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.676120 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-0\") pod \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.676166 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-telemetry-power-monitoring-combined-ca-bundle\") pod \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.676232 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-inventory\") pod \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.676281 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-2\") pod \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\" (UID: \"dfdfd706-a8f9-49ab-aba5-514d80a90c96\") " Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.683822 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfdfd706-a8f9-49ab-aba5-514d80a90c96-kube-api-access-j7442" (OuterVolumeSpecName: "kube-api-access-j7442") pod "dfdfd706-a8f9-49ab-aba5-514d80a90c96" (UID: "dfdfd706-a8f9-49ab-aba5-514d80a90c96"). InnerVolumeSpecName "kube-api-access-j7442". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.683834 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "dfdfd706-a8f9-49ab-aba5-514d80a90c96" (UID: "dfdfd706-a8f9-49ab-aba5-514d80a90c96"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.711263 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-inventory" (OuterVolumeSpecName: "inventory") pod "dfdfd706-a8f9-49ab-aba5-514d80a90c96" (UID: "dfdfd706-a8f9-49ab-aba5-514d80a90c96"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.711875 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "dfdfd706-a8f9-49ab-aba5-514d80a90c96" (UID: "dfdfd706-a8f9-49ab-aba5-514d80a90c96"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.716099 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "dfdfd706-a8f9-49ab-aba5-514d80a90c96" (UID: "dfdfd706-a8f9-49ab-aba5-514d80a90c96"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.716890 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "dfdfd706-a8f9-49ab-aba5-514d80a90c96" (UID: "dfdfd706-a8f9-49ab-aba5-514d80a90c96"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.717333 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "dfdfd706-a8f9-49ab-aba5-514d80a90c96" (UID: "dfdfd706-a8f9-49ab-aba5-514d80a90c96"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.778402 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j7442\" (UniqueName: \"kubernetes.io/projected/dfdfd706-a8f9-49ab-aba5-514d80a90c96-kube-api-access-j7442\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.778433 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.778443 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.778495 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.778509 4787 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.778544 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:38 crc kubenswrapper[4787]: I1203 18:00:38.778554 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/dfdfd706-a8f9-49ab-aba5-514d80a90c96-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.140942 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" event={"ID":"dfdfd706-a8f9-49ab-aba5-514d80a90c96","Type":"ContainerDied","Data":"f7283604c64a009d71276a4521622252e09ac1d94a91826a07d014dd818b62b5"} Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.141012 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f7283604c64a009d71276a4521622252e09ac1d94a91826a07d014dd818b62b5" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.141059 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.259577 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9"] Dec 03 18:00:39 crc kubenswrapper[4787]: E1203 18:00:39.260216 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfdfd706-a8f9-49ab-aba5-514d80a90c96" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.260246 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfdfd706-a8f9-49ab-aba5-514d80a90c96" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 18:00:39 crc kubenswrapper[4787]: E1203 18:00:39.260276 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f03a6f41-66de-4c21-84a1-8c30b66c80b8" containerName="collect-profiles" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.260288 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f03a6f41-66de-4c21-84a1-8c30b66c80b8" containerName="collect-profiles" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.260662 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f03a6f41-66de-4c21-84a1-8c30b66c80b8" containerName="collect-profiles" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.260704 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfdfd706-a8f9-49ab-aba5-514d80a90c96" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.261661 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.263974 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.264669 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.267124 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.267235 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.267547 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.282702 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9"] Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.389739 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.389809 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.390050 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqk8g\" (UniqueName: \"kubernetes.io/projected/073fd174-87b7-48f4-82a0-107986d6d72d-kube-api-access-kqk8g\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.390313 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.390517 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.492222 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.492322 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.492480 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.492525 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.492591 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqk8g\" (UniqueName: \"kubernetes.io/projected/073fd174-87b7-48f4-82a0-107986d6d72d-kube-api-access-kqk8g\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.498048 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.498521 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.499105 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.499471 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.515803 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqk8g\" (UniqueName: \"kubernetes.io/projected/073fd174-87b7-48f4-82a0-107986d6d72d-kube-api-access-kqk8g\") pod \"logging-edpm-deployment-openstack-edpm-ipam-cw6s9\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:39 crc kubenswrapper[4787]: I1203 18:00:39.582518 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:00:40 crc kubenswrapper[4787]: I1203 18:00:40.179035 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9"] Dec 03 18:00:40 crc kubenswrapper[4787]: I1203 18:00:40.181155 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:00:41 crc kubenswrapper[4787]: I1203 18:00:41.161828 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" event={"ID":"073fd174-87b7-48f4-82a0-107986d6d72d","Type":"ContainerStarted","Data":"dab7c1a7af8651abab7893e3de4bdf578dd9e6eb24ecaa656615c1042a292c03"} Dec 03 18:00:41 crc kubenswrapper[4787]: I1203 18:00:41.162135 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" event={"ID":"073fd174-87b7-48f4-82a0-107986d6d72d","Type":"ContainerStarted","Data":"4c79db4347b24e455ac9af839e8753aeeea0d281479c21640af7691157e93296"} Dec 03 18:00:41 crc kubenswrapper[4787]: I1203 18:00:41.194921 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" podStartSLOduration=1.729568 podStartE2EDuration="2.194898673s" podCreationTimestamp="2025-12-03 18:00:39 +0000 UTC" firstStartedPulling="2025-12-03 18:00:40.180868334 +0000 UTC m=+2856.998339313" lastFinishedPulling="2025-12-03 18:00:40.646198987 +0000 UTC m=+2857.463669986" observedRunningTime="2025-12-03 18:00:41.178900255 +0000 UTC m=+2857.996371234" watchObservedRunningTime="2025-12-03 18:00:41.194898673 +0000 UTC m=+2858.012369652" Dec 03 18:00:48 crc kubenswrapper[4787]: E1203 18:00:48.316940 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice/crio-c27babdd468ed5ac6e4ee2b5e121475f105d11c580cc49879f629dabad408d2e\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice\": RecentStats: unable to find data in memory cache]" Dec 03 18:00:48 crc kubenswrapper[4787]: I1203 18:00:48.989756 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:00:48 crc kubenswrapper[4787]: I1203 18:00:48.989811 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:00:58 crc kubenswrapper[4787]: E1203 18:00:58.616861 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf03a6f41_66de_4c21_84a1_8c30b66c80b8.slice/crio-c27babdd468ed5ac6e4ee2b5e121475f105d11c580cc49879f629dabad408d2e\": RecentStats: unable to find data in memory cache]" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.157913 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29413081-xcmrh"] Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.161006 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.177130 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413081-xcmrh"] Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.258612 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n9nvd\" (UniqueName: \"kubernetes.io/projected/762a8bc2-3aa0-498a-852a-3ede435f23df-kube-api-access-n9nvd\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.258815 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-combined-ca-bundle\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.258890 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-fernet-keys\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.259228 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-config-data\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.361953 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-config-data\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.362147 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n9nvd\" (UniqueName: \"kubernetes.io/projected/762a8bc2-3aa0-498a-852a-3ede435f23df-kube-api-access-n9nvd\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.362239 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-combined-ca-bundle\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.362293 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-fernet-keys\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.370059 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-config-data\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.370977 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-fernet-keys\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.371521 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-combined-ca-bundle\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.385675 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n9nvd\" (UniqueName: \"kubernetes.io/projected/762a8bc2-3aa0-498a-852a-3ede435f23df-kube-api-access-n9nvd\") pod \"keystone-cron-29413081-xcmrh\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:00 crc kubenswrapper[4787]: I1203 18:01:00.496593 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:01 crc kubenswrapper[4787]: I1203 18:01:01.405585 4787 generic.go:334] "Generic (PLEG): container finished" podID="073fd174-87b7-48f4-82a0-107986d6d72d" containerID="dab7c1a7af8651abab7893e3de4bdf578dd9e6eb24ecaa656615c1042a292c03" exitCode=0 Dec 03 18:01:01 crc kubenswrapper[4787]: I1203 18:01:01.405666 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" event={"ID":"073fd174-87b7-48f4-82a0-107986d6d72d","Type":"ContainerDied","Data":"dab7c1a7af8651abab7893e3de4bdf578dd9e6eb24ecaa656615c1042a292c03"} Dec 03 18:01:01 crc kubenswrapper[4787]: I1203 18:01:01.555386 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413081-xcmrh"] Dec 03 18:01:01 crc kubenswrapper[4787]: W1203 18:01:01.559806 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod762a8bc2_3aa0_498a_852a_3ede435f23df.slice/crio-9b0cad42799156acd8d86eec585bb438bf4faff1a81761326cab0d798d722092 WatchSource:0}: Error finding container 9b0cad42799156acd8d86eec585bb438bf4faff1a81761326cab0d798d722092: Status 404 returned error can't find the container with id 9b0cad42799156acd8d86eec585bb438bf4faff1a81761326cab0d798d722092 Dec 03 18:01:02 crc kubenswrapper[4787]: I1203 18:01:02.435168 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413081-xcmrh" event={"ID":"762a8bc2-3aa0-498a-852a-3ede435f23df","Type":"ContainerStarted","Data":"746f5439cf927c9f099b1f6e6bdd8fb35b40ac18452daec78d23288470fc236e"} Dec 03 18:01:02 crc kubenswrapper[4787]: I1203 18:01:02.435507 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413081-xcmrh" event={"ID":"762a8bc2-3aa0-498a-852a-3ede435f23df","Type":"ContainerStarted","Data":"9b0cad42799156acd8d86eec585bb438bf4faff1a81761326cab0d798d722092"} Dec 03 18:01:02 crc kubenswrapper[4787]: I1203 18:01:02.467709 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29413081-xcmrh" podStartSLOduration=2.467686047 podStartE2EDuration="2.467686047s" podCreationTimestamp="2025-12-03 18:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:01:02.452084669 +0000 UTC m=+2879.269555648" watchObservedRunningTime="2025-12-03 18:01:02.467686047 +0000 UTC m=+2879.285157016" Dec 03 18:01:02 crc kubenswrapper[4787]: I1203 18:01:02.902579 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.023512 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-ssh-key\") pod \"073fd174-87b7-48f4-82a0-107986d6d72d\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.023690 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-0\") pod \"073fd174-87b7-48f4-82a0-107986d6d72d\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.023821 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-1\") pod \"073fd174-87b7-48f4-82a0-107986d6d72d\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.023932 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-inventory\") pod \"073fd174-87b7-48f4-82a0-107986d6d72d\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.023958 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqk8g\" (UniqueName: \"kubernetes.io/projected/073fd174-87b7-48f4-82a0-107986d6d72d-kube-api-access-kqk8g\") pod \"073fd174-87b7-48f4-82a0-107986d6d72d\" (UID: \"073fd174-87b7-48f4-82a0-107986d6d72d\") " Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.030798 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/073fd174-87b7-48f4-82a0-107986d6d72d-kube-api-access-kqk8g" (OuterVolumeSpecName: "kube-api-access-kqk8g") pod "073fd174-87b7-48f4-82a0-107986d6d72d" (UID: "073fd174-87b7-48f4-82a0-107986d6d72d"). InnerVolumeSpecName "kube-api-access-kqk8g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.059709 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "073fd174-87b7-48f4-82a0-107986d6d72d" (UID: "073fd174-87b7-48f4-82a0-107986d6d72d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.062703 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-inventory" (OuterVolumeSpecName: "inventory") pod "073fd174-87b7-48f4-82a0-107986d6d72d" (UID: "073fd174-87b7-48f4-82a0-107986d6d72d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.063129 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "073fd174-87b7-48f4-82a0-107986d6d72d" (UID: "073fd174-87b7-48f4-82a0-107986d6d72d"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.066409 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "073fd174-87b7-48f4-82a0-107986d6d72d" (UID: "073fd174-87b7-48f4-82a0-107986d6d72d"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.126740 4787 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.127094 4787 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.127117 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.127136 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqk8g\" (UniqueName: \"kubernetes.io/projected/073fd174-87b7-48f4-82a0-107986d6d72d-kube-api-access-kqk8g\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.127154 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/073fd174-87b7-48f4-82a0-107986d6d72d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.457672 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.457698 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9" event={"ID":"073fd174-87b7-48f4-82a0-107986d6d72d","Type":"ContainerDied","Data":"4c79db4347b24e455ac9af839e8753aeeea0d281479c21640af7691157e93296"} Dec 03 18:01:03 crc kubenswrapper[4787]: I1203 18:01:03.457757 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c79db4347b24e455ac9af839e8753aeeea0d281479c21640af7691157e93296" Dec 03 18:01:04 crc kubenswrapper[4787]: I1203 18:01:04.468827 4787 generic.go:334] "Generic (PLEG): container finished" podID="762a8bc2-3aa0-498a-852a-3ede435f23df" containerID="746f5439cf927c9f099b1f6e6bdd8fb35b40ac18452daec78d23288470fc236e" exitCode=0 Dec 03 18:01:04 crc kubenswrapper[4787]: I1203 18:01:04.468925 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413081-xcmrh" event={"ID":"762a8bc2-3aa0-498a-852a-3ede435f23df","Type":"ContainerDied","Data":"746f5439cf927c9f099b1f6e6bdd8fb35b40ac18452daec78d23288470fc236e"} Dec 03 18:01:05 crc kubenswrapper[4787]: I1203 18:01:05.887057 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:05 crc kubenswrapper[4787]: I1203 18:01:05.996319 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-config-data\") pod \"762a8bc2-3aa0-498a-852a-3ede435f23df\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " Dec 03 18:01:05 crc kubenswrapper[4787]: I1203 18:01:05.996636 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-n9nvd\" (UniqueName: \"kubernetes.io/projected/762a8bc2-3aa0-498a-852a-3ede435f23df-kube-api-access-n9nvd\") pod \"762a8bc2-3aa0-498a-852a-3ede435f23df\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " Dec 03 18:01:05 crc kubenswrapper[4787]: I1203 18:01:05.996765 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-combined-ca-bundle\") pod \"762a8bc2-3aa0-498a-852a-3ede435f23df\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " Dec 03 18:01:05 crc kubenswrapper[4787]: I1203 18:01:05.996914 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-fernet-keys\") pod \"762a8bc2-3aa0-498a-852a-3ede435f23df\" (UID: \"762a8bc2-3aa0-498a-852a-3ede435f23df\") " Dec 03 18:01:06 crc kubenswrapper[4787]: I1203 18:01:06.002476 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "762a8bc2-3aa0-498a-852a-3ede435f23df" (UID: "762a8bc2-3aa0-498a-852a-3ede435f23df"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:01:06 crc kubenswrapper[4787]: I1203 18:01:06.002896 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/762a8bc2-3aa0-498a-852a-3ede435f23df-kube-api-access-n9nvd" (OuterVolumeSpecName: "kube-api-access-n9nvd") pod "762a8bc2-3aa0-498a-852a-3ede435f23df" (UID: "762a8bc2-3aa0-498a-852a-3ede435f23df"). InnerVolumeSpecName "kube-api-access-n9nvd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:01:06 crc kubenswrapper[4787]: I1203 18:01:06.049315 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "762a8bc2-3aa0-498a-852a-3ede435f23df" (UID: "762a8bc2-3aa0-498a-852a-3ede435f23df"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:01:06 crc kubenswrapper[4787]: I1203 18:01:06.094418 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-config-data" (OuterVolumeSpecName: "config-data") pod "762a8bc2-3aa0-498a-852a-3ede435f23df" (UID: "762a8bc2-3aa0-498a-852a-3ede435f23df"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:01:06 crc kubenswrapper[4787]: I1203 18:01:06.101980 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-n9nvd\" (UniqueName: \"kubernetes.io/projected/762a8bc2-3aa0-498a-852a-3ede435f23df-kube-api-access-n9nvd\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:06 crc kubenswrapper[4787]: I1203 18:01:06.102204 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:06 crc kubenswrapper[4787]: I1203 18:01:06.102238 4787 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:06 crc kubenswrapper[4787]: I1203 18:01:06.102308 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/762a8bc2-3aa0-498a-852a-3ede435f23df-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:01:06 crc kubenswrapper[4787]: I1203 18:01:06.492525 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413081-xcmrh" event={"ID":"762a8bc2-3aa0-498a-852a-3ede435f23df","Type":"ContainerDied","Data":"9b0cad42799156acd8d86eec585bb438bf4faff1a81761326cab0d798d722092"} Dec 03 18:01:06 crc kubenswrapper[4787]: I1203 18:01:06.492565 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b0cad42799156acd8d86eec585bb438bf4faff1a81761326cab0d798d722092" Dec 03 18:01:06 crc kubenswrapper[4787]: I1203 18:01:06.492813 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413081-xcmrh" Dec 03 18:01:18 crc kubenswrapper[4787]: I1203 18:01:18.989525 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:01:18 crc kubenswrapper[4787]: I1203 18:01:18.990158 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:01:48 crc kubenswrapper[4787]: I1203 18:01:48.989624 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:01:48 crc kubenswrapper[4787]: I1203 18:01:48.990332 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:01:48 crc kubenswrapper[4787]: I1203 18:01:48.990546 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 18:01:48 crc kubenswrapper[4787]: I1203 18:01:48.991695 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5a433e35c83e9c38d50149fbfb3dd5416667750d83bba18e735da69f228bbfe5"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:01:48 crc kubenswrapper[4787]: I1203 18:01:48.991789 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://5a433e35c83e9c38d50149fbfb3dd5416667750d83bba18e735da69f228bbfe5" gracePeriod=600 Dec 03 18:01:50 crc kubenswrapper[4787]: I1203 18:01:50.007217 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="5a433e35c83e9c38d50149fbfb3dd5416667750d83bba18e735da69f228bbfe5" exitCode=0 Dec 03 18:01:50 crc kubenswrapper[4787]: I1203 18:01:50.007290 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"5a433e35c83e9c38d50149fbfb3dd5416667750d83bba18e735da69f228bbfe5"} Dec 03 18:01:50 crc kubenswrapper[4787]: I1203 18:01:50.007495 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8"} Dec 03 18:01:50 crc kubenswrapper[4787]: I1203 18:01:50.007519 4787 scope.go:117] "RemoveContainer" containerID="5290b5bef48dde3006be660e3b63d544cf120231425b9cfaae14e1761c3f01d9" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.000952 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-568xk"] Dec 03 18:03:52 crc kubenswrapper[4787]: E1203 18:03:52.002436 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="762a8bc2-3aa0-498a-852a-3ede435f23df" containerName="keystone-cron" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.002461 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="762a8bc2-3aa0-498a-852a-3ede435f23df" containerName="keystone-cron" Dec 03 18:03:52 crc kubenswrapper[4787]: E1203 18:03:52.002520 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="073fd174-87b7-48f4-82a0-107986d6d72d" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.002537 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="073fd174-87b7-48f4-82a0-107986d6d72d" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.002971 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="073fd174-87b7-48f4-82a0-107986d6d72d" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.003056 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="762a8bc2-3aa0-498a-852a-3ede435f23df" containerName="keystone-cron" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.006730 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-568xk" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.029961 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-568xk"] Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.085198 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-utilities\") pod \"community-operators-568xk\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " pod="openshift-marketplace/community-operators-568xk" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.085428 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-catalog-content\") pod \"community-operators-568xk\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " pod="openshift-marketplace/community-operators-568xk" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.085845 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nnxnk\" (UniqueName: \"kubernetes.io/projected/a88bb365-0e79-463c-a4c1-8c2513560e44-kube-api-access-nnxnk\") pod \"community-operators-568xk\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " pod="openshift-marketplace/community-operators-568xk" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.187504 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-utilities\") pod \"community-operators-568xk\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " pod="openshift-marketplace/community-operators-568xk" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.187605 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-catalog-content\") pod \"community-operators-568xk\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " pod="openshift-marketplace/community-operators-568xk" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.187722 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nnxnk\" (UniqueName: \"kubernetes.io/projected/a88bb365-0e79-463c-a4c1-8c2513560e44-kube-api-access-nnxnk\") pod \"community-operators-568xk\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " pod="openshift-marketplace/community-operators-568xk" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.187946 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-utilities\") pod \"community-operators-568xk\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " pod="openshift-marketplace/community-operators-568xk" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.188049 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-catalog-content\") pod \"community-operators-568xk\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " pod="openshift-marketplace/community-operators-568xk" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.209035 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nnxnk\" (UniqueName: \"kubernetes.io/projected/a88bb365-0e79-463c-a4c1-8c2513560e44-kube-api-access-nnxnk\") pod \"community-operators-568xk\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " pod="openshift-marketplace/community-operators-568xk" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.330972 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-568xk" Dec 03 18:03:52 crc kubenswrapper[4787]: I1203 18:03:52.904463 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-568xk"] Dec 03 18:03:53 crc kubenswrapper[4787]: I1203 18:03:53.430511 4787 generic.go:334] "Generic (PLEG): container finished" podID="a88bb365-0e79-463c-a4c1-8c2513560e44" containerID="8d600f2f5e04e4b67e4105be4960542fe5ac47e1133dc109de4575fd0654a391" exitCode=0 Dec 03 18:03:53 crc kubenswrapper[4787]: I1203 18:03:53.430851 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-568xk" event={"ID":"a88bb365-0e79-463c-a4c1-8c2513560e44","Type":"ContainerDied","Data":"8d600f2f5e04e4b67e4105be4960542fe5ac47e1133dc109de4575fd0654a391"} Dec 03 18:03:53 crc kubenswrapper[4787]: I1203 18:03:53.430894 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-568xk" event={"ID":"a88bb365-0e79-463c-a4c1-8c2513560e44","Type":"ContainerStarted","Data":"81f0e18de2ef8a003ed670a5c46d76aec09c9f9708dcc745458b7487d78d86f9"} Dec 03 18:03:54 crc kubenswrapper[4787]: I1203 18:03:54.446637 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-568xk" event={"ID":"a88bb365-0e79-463c-a4c1-8c2513560e44","Type":"ContainerStarted","Data":"e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782"} Dec 03 18:03:55 crc kubenswrapper[4787]: I1203 18:03:55.465355 4787 generic.go:334] "Generic (PLEG): container finished" podID="a88bb365-0e79-463c-a4c1-8c2513560e44" containerID="e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782" exitCode=0 Dec 03 18:03:55 crc kubenswrapper[4787]: I1203 18:03:55.465486 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-568xk" event={"ID":"a88bb365-0e79-463c-a4c1-8c2513560e44","Type":"ContainerDied","Data":"e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782"} Dec 03 18:03:57 crc kubenswrapper[4787]: I1203 18:03:57.486639 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-568xk" event={"ID":"a88bb365-0e79-463c-a4c1-8c2513560e44","Type":"ContainerStarted","Data":"86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60"} Dec 03 18:03:57 crc kubenswrapper[4787]: I1203 18:03:57.515903 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-568xk" podStartSLOduration=3.354979722 podStartE2EDuration="6.515877989s" podCreationTimestamp="2025-12-03 18:03:51 +0000 UTC" firstStartedPulling="2025-12-03 18:03:53.434500708 +0000 UTC m=+3050.251971657" lastFinishedPulling="2025-12-03 18:03:56.595398965 +0000 UTC m=+3053.412869924" observedRunningTime="2025-12-03 18:03:57.50398548 +0000 UTC m=+3054.321456459" watchObservedRunningTime="2025-12-03 18:03:57.515877989 +0000 UTC m=+3054.333348938" Dec 03 18:04:02 crc kubenswrapper[4787]: I1203 18:04:02.331129 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-568xk" Dec 03 18:04:02 crc kubenswrapper[4787]: I1203 18:04:02.331748 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-568xk" Dec 03 18:04:02 crc kubenswrapper[4787]: I1203 18:04:02.388960 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-568xk" Dec 03 18:04:02 crc kubenswrapper[4787]: I1203 18:04:02.589414 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-568xk" Dec 03 18:04:05 crc kubenswrapper[4787]: I1203 18:04:05.378041 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-568xk"] Dec 03 18:04:05 crc kubenswrapper[4787]: I1203 18:04:05.378867 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-568xk" podUID="a88bb365-0e79-463c-a4c1-8c2513560e44" containerName="registry-server" containerID="cri-o://86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60" gracePeriod=2 Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.396168 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-568xk" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.531733 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-utilities\") pod \"a88bb365-0e79-463c-a4c1-8c2513560e44\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.531897 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nnxnk\" (UniqueName: \"kubernetes.io/projected/a88bb365-0e79-463c-a4c1-8c2513560e44-kube-api-access-nnxnk\") pod \"a88bb365-0e79-463c-a4c1-8c2513560e44\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.532054 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-catalog-content\") pod \"a88bb365-0e79-463c-a4c1-8c2513560e44\" (UID: \"a88bb365-0e79-463c-a4c1-8c2513560e44\") " Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.541306 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-utilities" (OuterVolumeSpecName: "utilities") pod "a88bb365-0e79-463c-a4c1-8c2513560e44" (UID: "a88bb365-0e79-463c-a4c1-8c2513560e44"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.545483 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a88bb365-0e79-463c-a4c1-8c2513560e44-kube-api-access-nnxnk" (OuterVolumeSpecName: "kube-api-access-nnxnk") pod "a88bb365-0e79-463c-a4c1-8c2513560e44" (UID: "a88bb365-0e79-463c-a4c1-8c2513560e44"). InnerVolumeSpecName "kube-api-access-nnxnk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.581318 4787 generic.go:334] "Generic (PLEG): container finished" podID="a88bb365-0e79-463c-a4c1-8c2513560e44" containerID="86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60" exitCode=0 Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.581365 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-568xk" event={"ID":"a88bb365-0e79-463c-a4c1-8c2513560e44","Type":"ContainerDied","Data":"86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60"} Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.581391 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-568xk" event={"ID":"a88bb365-0e79-463c-a4c1-8c2513560e44","Type":"ContainerDied","Data":"81f0e18de2ef8a003ed670a5c46d76aec09c9f9708dcc745458b7487d78d86f9"} Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.581407 4787 scope.go:117] "RemoveContainer" containerID="86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.582591 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-568xk" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.590652 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a88bb365-0e79-463c-a4c1-8c2513560e44" (UID: "a88bb365-0e79-463c-a4c1-8c2513560e44"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.623659 4787 scope.go:117] "RemoveContainer" containerID="e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.634338 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.634377 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nnxnk\" (UniqueName: \"kubernetes.io/projected/a88bb365-0e79-463c-a4c1-8c2513560e44-kube-api-access-nnxnk\") on node \"crc\" DevicePath \"\"" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.634417 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a88bb365-0e79-463c-a4c1-8c2513560e44-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.658402 4787 scope.go:117] "RemoveContainer" containerID="8d600f2f5e04e4b67e4105be4960542fe5ac47e1133dc109de4575fd0654a391" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.721700 4787 scope.go:117] "RemoveContainer" containerID="86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60" Dec 03 18:04:06 crc kubenswrapper[4787]: E1203 18:04:06.722249 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60\": container with ID starting with 86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60 not found: ID does not exist" containerID="86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.722276 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60"} err="failed to get container status \"86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60\": rpc error: code = NotFound desc = could not find container \"86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60\": container with ID starting with 86696c312c33612e1447988c993d83d5862dd187105106255764c5d643e12d60 not found: ID does not exist" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.722294 4787 scope.go:117] "RemoveContainer" containerID="e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782" Dec 03 18:04:06 crc kubenswrapper[4787]: E1203 18:04:06.722689 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782\": container with ID starting with e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782 not found: ID does not exist" containerID="e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.722733 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782"} err="failed to get container status \"e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782\": rpc error: code = NotFound desc = could not find container \"e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782\": container with ID starting with e742715274509c54b9a9ace91172ac28ebef47257897d41bed09dd1a8dcf9782 not found: ID does not exist" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.722760 4787 scope.go:117] "RemoveContainer" containerID="8d600f2f5e04e4b67e4105be4960542fe5ac47e1133dc109de4575fd0654a391" Dec 03 18:04:06 crc kubenswrapper[4787]: E1203 18:04:06.723144 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8d600f2f5e04e4b67e4105be4960542fe5ac47e1133dc109de4575fd0654a391\": container with ID starting with 8d600f2f5e04e4b67e4105be4960542fe5ac47e1133dc109de4575fd0654a391 not found: ID does not exist" containerID="8d600f2f5e04e4b67e4105be4960542fe5ac47e1133dc109de4575fd0654a391" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.723175 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8d600f2f5e04e4b67e4105be4960542fe5ac47e1133dc109de4575fd0654a391"} err="failed to get container status \"8d600f2f5e04e4b67e4105be4960542fe5ac47e1133dc109de4575fd0654a391\": rpc error: code = NotFound desc = could not find container \"8d600f2f5e04e4b67e4105be4960542fe5ac47e1133dc109de4575fd0654a391\": container with ID starting with 8d600f2f5e04e4b67e4105be4960542fe5ac47e1133dc109de4575fd0654a391 not found: ID does not exist" Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.919367 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-568xk"] Dec 03 18:04:06 crc kubenswrapper[4787]: I1203 18:04:06.929310 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-568xk"] Dec 03 18:04:07 crc kubenswrapper[4787]: I1203 18:04:07.780437 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a88bb365-0e79-463c-a4c1-8c2513560e44" path="/var/lib/kubelet/pods/a88bb365-0e79-463c-a4c1-8c2513560e44/volumes" Dec 03 18:04:18 crc kubenswrapper[4787]: I1203 18:04:18.989817 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:04:18 crc kubenswrapper[4787]: I1203 18:04:18.990365 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:04:48 crc kubenswrapper[4787]: I1203 18:04:48.990339 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:04:48 crc kubenswrapper[4787]: I1203 18:04:48.990746 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:05:18 crc kubenswrapper[4787]: I1203 18:05:18.989492 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:05:18 crc kubenswrapper[4787]: I1203 18:05:18.990263 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:05:18 crc kubenswrapper[4787]: I1203 18:05:18.990348 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 18:05:18 crc kubenswrapper[4787]: I1203 18:05:18.991605 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:05:18 crc kubenswrapper[4787]: I1203 18:05:18.991702 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" gracePeriod=600 Dec 03 18:05:19 crc kubenswrapper[4787]: E1203 18:05:19.670126 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:05:20 crc kubenswrapper[4787]: I1203 18:05:20.367751 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" exitCode=0 Dec 03 18:05:20 crc kubenswrapper[4787]: I1203 18:05:20.367787 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8"} Dec 03 18:05:20 crc kubenswrapper[4787]: I1203 18:05:20.368079 4787 scope.go:117] "RemoveContainer" containerID="5a433e35c83e9c38d50149fbfb3dd5416667750d83bba18e735da69f228bbfe5" Dec 03 18:05:20 crc kubenswrapper[4787]: I1203 18:05:20.371199 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:05:20 crc kubenswrapper[4787]: E1203 18:05:20.371589 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:05:31 crc kubenswrapper[4787]: I1203 18:05:31.766008 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:05:31 crc kubenswrapper[4787]: E1203 18:05:31.766796 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:05:44 crc kubenswrapper[4787]: I1203 18:05:44.767182 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:05:44 crc kubenswrapper[4787]: E1203 18:05:44.768227 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:05:58 crc kubenswrapper[4787]: I1203 18:05:58.767298 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:05:58 crc kubenswrapper[4787]: E1203 18:05:58.768802 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:06:13 crc kubenswrapper[4787]: I1203 18:06:13.772823 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:06:13 crc kubenswrapper[4787]: E1203 18:06:13.773616 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.685582 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.699560 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-rtc69"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.722225 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.735059 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.749490 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.757919 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.778944 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2aabe7bb-06d8-44b6-a293-d228d461c1ab" path="/var/lib/kubelet/pods/2aabe7bb-06d8-44b6-a293-d228d461c1ab/volumes" Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.779829 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.779860 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.783373 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.791750 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-jjskt"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.799570 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-xbpsk"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.807162 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-z9phj"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.814753 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.823079 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.831059 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-nm2hf"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.839470 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-kpr56"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.846762 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-lbl9g"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.854710 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.863070 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-cw6s9"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.872079 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-g975p"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.879598 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.887699 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.895830 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-7x6rk"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.903957 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-s4chd"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.912142 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-lvvx2"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.920441 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-zsjqn"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.930514 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.936422 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-z9phj"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.953104 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.959051 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-r5sh2"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.971092 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-ltz6d"] Dec 03 18:06:19 crc kubenswrapper[4787]: I1203 18:06:19.980087 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-dql8f"] Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.781843 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="073fd174-87b7-48f4-82a0-107986d6d72d" path="/var/lib/kubelet/pods/073fd174-87b7-48f4-82a0-107986d6d72d/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.783198 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="251ef657-822b-4392-9c92-266ccd5ca952" path="/var/lib/kubelet/pods/251ef657-822b-4392-9c92-266ccd5ca952/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.784051 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ce09f0a-e6a4-4b04-a686-d43e1882f358" path="/var/lib/kubelet/pods/5ce09f0a-e6a4-4b04-a686-d43e1882f358/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.784883 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d4c7463-edc8-40f9-80f6-65fe7d4e40b8" path="/var/lib/kubelet/pods/5d4c7463-edc8-40f9-80f6-65fe7d4e40b8/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.786429 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="66d0ee2f-607b-4a9c-9808-4200fef6c906" path="/var/lib/kubelet/pods/66d0ee2f-607b-4a9c-9808-4200fef6c906/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.787271 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6aaee49b-7923-4882-9db0-28ad5d07dea9" path="/var/lib/kubelet/pods/6aaee49b-7923-4882-9db0-28ad5d07dea9/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.788240 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b" path="/var/lib/kubelet/pods/99e1ef0d-6d9c-413a-9d70-e8b0aa3f441b/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.789656 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb00d591-6e35-41bd-be71-a8a8190946b1" path="/var/lib/kubelet/pods/bb00d591-6e35-41bd-be71-a8a8190946b1/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.790491 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c408e255-873c-4200-9768-6576ea6f6b0c" path="/var/lib/kubelet/pods/c408e255-873c-4200-9768-6576ea6f6b0c/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.791276 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08" path="/var/lib/kubelet/pods/dbaf8a7d-b17e-4a83-a754-cec5ff3d6a08/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.792785 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfdfd706-a8f9-49ab-aba5-514d80a90c96" path="/var/lib/kubelet/pods/dfdfd706-a8f9-49ab-aba5-514d80a90c96/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.793562 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0400992-8418-4339-9162-f788984931e1" path="/var/lib/kubelet/pods/e0400992-8418-4339-9162-f788984931e1/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.794351 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e3e4d4bd-145d-417d-a792-b74eed16e8e0" path="/var/lib/kubelet/pods/e3e4d4bd-145d-417d-a792-b74eed16e8e0/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.795100 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f30ff726-5f48-483b-a5e4-7f49050157f6" path="/var/lib/kubelet/pods/f30ff726-5f48-483b-a5e4-7f49050157f6/volumes" Dec 03 18:06:21 crc kubenswrapper[4787]: I1203 18:06:21.796432 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f34ffd31-6238-4f68-80cc-f07be91a44f8" path="/var/lib/kubelet/pods/f34ffd31-6238-4f68-80cc-f07be91a44f8/volumes" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.685523 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd"] Dec 03 18:06:24 crc kubenswrapper[4787]: E1203 18:06:24.686432 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a88bb365-0e79-463c-a4c1-8c2513560e44" containerName="extract-content" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.686444 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88bb365-0e79-463c-a4c1-8c2513560e44" containerName="extract-content" Dec 03 18:06:24 crc kubenswrapper[4787]: E1203 18:06:24.686461 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a88bb365-0e79-463c-a4c1-8c2513560e44" containerName="extract-utilities" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.686468 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88bb365-0e79-463c-a4c1-8c2513560e44" containerName="extract-utilities" Dec 03 18:06:24 crc kubenswrapper[4787]: E1203 18:06:24.686483 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a88bb365-0e79-463c-a4c1-8c2513560e44" containerName="registry-server" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.686490 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a88bb365-0e79-463c-a4c1-8c2513560e44" containerName="registry-server" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.686693 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a88bb365-0e79-463c-a4c1-8c2513560e44" containerName="registry-server" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.687396 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.694632 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.694669 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.694743 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.694837 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.694851 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.698102 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd"] Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.766577 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:06:24 crc kubenswrapper[4787]: E1203 18:06:24.766878 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.785080 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.785165 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwjrz\" (UniqueName: \"kubernetes.io/projected/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-kube-api-access-hwjrz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.785355 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.785382 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.785546 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.887949 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.888007 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.888180 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.888304 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.888336 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwjrz\" (UniqueName: \"kubernetes.io/projected/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-kube-api-access-hwjrz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.894725 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.895606 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.896147 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.896324 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:24 crc kubenswrapper[4787]: I1203 18:06:24.904925 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwjrz\" (UniqueName: \"kubernetes.io/projected/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-kube-api-access-hwjrz\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.017127 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.489205 4787 scope.go:117] "RemoveContainer" containerID="ed0e7cff92f75edd35950004a414ef9da7a31f4b1eaa2bccd6b52cd682582c65" Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.533795 4787 scope.go:117] "RemoveContainer" containerID="3b79134240d56cbb27fe03ad1a79b5e97a9eb8fc8fe6f48d84c0f14dcd94d033" Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.600362 4787 scope.go:117] "RemoveContainer" containerID="1667e55d178e8f38a692c5ea94bde08a6200d42064af14191e9ac90e82e74946" Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.612001 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd"] Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.647734 4787 scope.go:117] "RemoveContainer" containerID="74b0cc42250f3c808d019102b9ee07c7623bbfd7ac34b6181b9ff01c02036a9e" Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.648294 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.682768 4787 scope.go:117] "RemoveContainer" containerID="762d7ae422f90cd7ab8b2856d50541644f647d79c9c4d20957c45e0fbf9075ef" Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.760538 4787 scope.go:117] "RemoveContainer" containerID="9b616255df00e40b54d812bd65dc8d85465f9f84eb0b597a78083c6c41152ece" Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.803611 4787 scope.go:117] "RemoveContainer" containerID="506f9832b7028efce81cadd12518b375f1a755602fdc73cfddeb622111b0cbc4" Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.853680 4787 scope.go:117] "RemoveContainer" containerID="c67efae0d5b9257e3f2d0e69bc7e4a0d411c7e3f666e0e644080c84fcc472d71" Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.883310 4787 scope.go:117] "RemoveContainer" containerID="a82102220a8491ba74af32d29889eaab4a9504dec92e3b3f7feec792ca5b8c0e" Dec 03 18:06:25 crc kubenswrapper[4787]: I1203 18:06:25.951076 4787 scope.go:117] "RemoveContainer" containerID="ad1416f9bce630e669d0ff1bc345a80417673ceb40c0f5973c46e096978e87b1" Dec 03 18:06:26 crc kubenswrapper[4787]: I1203 18:06:26.006675 4787 scope.go:117] "RemoveContainer" containerID="da01f1d732c8465b9ef4e94f3c6efd89b90a966451ccf642ee3263a38db1e7d8" Dec 03 18:06:26 crc kubenswrapper[4787]: I1203 18:06:26.067042 4787 scope.go:117] "RemoveContainer" containerID="52bac3d60506d1bb40b46024f3fbe02eb06329e1ec019f55628e2397d5d3062e" Dec 03 18:06:26 crc kubenswrapper[4787]: I1203 18:06:26.137325 4787 scope.go:117] "RemoveContainer" containerID="8dec503b458e8e8f0422ac38bb3eec52b51c3d02d46b7dafd0d516e3ef462128" Dec 03 18:06:26 crc kubenswrapper[4787]: I1203 18:06:26.139643 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" event={"ID":"be6c98e3-7b90-462f-ac9b-bf1874f97fc8","Type":"ContainerStarted","Data":"6e45bd11d3c380a8c96361741fcd3c0513a8f1ef2d26ef304a7e9cef23923db1"} Dec 03 18:06:26 crc kubenswrapper[4787]: I1203 18:06:26.198083 4787 scope.go:117] "RemoveContainer" containerID="7bf32c081a4b702652e0a6f9b5bc7ee04153f83fd34a4535b91f6ade01959fec" Dec 03 18:06:26 crc kubenswrapper[4787]: I1203 18:06:26.279955 4787 scope.go:117] "RemoveContainer" containerID="9365248c1aca69a164c2213cd1d1583fdcf9f1b00170793994c6a824c715ccad" Dec 03 18:06:27 crc kubenswrapper[4787]: I1203 18:06:27.177223 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" event={"ID":"be6c98e3-7b90-462f-ac9b-bf1874f97fc8","Type":"ContainerStarted","Data":"2164264b5b36d75907f1742664e8f06cbf568aa6011b6eff43207766e99d4b5f"} Dec 03 18:06:27 crc kubenswrapper[4787]: I1203 18:06:27.197286 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" podStartSLOduration=2.618144728 podStartE2EDuration="3.197259749s" podCreationTimestamp="2025-12-03 18:06:24 +0000 UTC" firstStartedPulling="2025-12-03 18:06:25.648072617 +0000 UTC m=+3202.465543576" lastFinishedPulling="2025-12-03 18:06:26.227187638 +0000 UTC m=+3203.044658597" observedRunningTime="2025-12-03 18:06:27.195576774 +0000 UTC m=+3204.013047763" watchObservedRunningTime="2025-12-03 18:06:27.197259749 +0000 UTC m=+3204.014730758" Dec 03 18:06:37 crc kubenswrapper[4787]: I1203 18:06:37.766357 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:06:37 crc kubenswrapper[4787]: E1203 18:06:37.767558 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:06:39 crc kubenswrapper[4787]: I1203 18:06:39.314240 4787 generic.go:334] "Generic (PLEG): container finished" podID="be6c98e3-7b90-462f-ac9b-bf1874f97fc8" containerID="2164264b5b36d75907f1742664e8f06cbf568aa6011b6eff43207766e99d4b5f" exitCode=0 Dec 03 18:06:39 crc kubenswrapper[4787]: I1203 18:06:39.314285 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" event={"ID":"be6c98e3-7b90-462f-ac9b-bf1874f97fc8","Type":"ContainerDied","Data":"2164264b5b36d75907f1742664e8f06cbf568aa6011b6eff43207766e99d4b5f"} Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.785464 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.851328 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ceph\") pod \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.851407 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-inventory\") pod \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.851439 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ssh-key\") pod \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.851471 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-repo-setup-combined-ca-bundle\") pod \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.851778 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwjrz\" (UniqueName: \"kubernetes.io/projected/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-kube-api-access-hwjrz\") pod \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\" (UID: \"be6c98e3-7b90-462f-ac9b-bf1874f97fc8\") " Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.860789 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "be6c98e3-7b90-462f-ac9b-bf1874f97fc8" (UID: "be6c98e3-7b90-462f-ac9b-bf1874f97fc8"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.860859 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ceph" (OuterVolumeSpecName: "ceph") pod "be6c98e3-7b90-462f-ac9b-bf1874f97fc8" (UID: "be6c98e3-7b90-462f-ac9b-bf1874f97fc8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.865098 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-kube-api-access-hwjrz" (OuterVolumeSpecName: "kube-api-access-hwjrz") pod "be6c98e3-7b90-462f-ac9b-bf1874f97fc8" (UID: "be6c98e3-7b90-462f-ac9b-bf1874f97fc8"). InnerVolumeSpecName "kube-api-access-hwjrz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.885590 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "be6c98e3-7b90-462f-ac9b-bf1874f97fc8" (UID: "be6c98e3-7b90-462f-ac9b-bf1874f97fc8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.892210 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-inventory" (OuterVolumeSpecName: "inventory") pod "be6c98e3-7b90-462f-ac9b-bf1874f97fc8" (UID: "be6c98e3-7b90-462f-ac9b-bf1874f97fc8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.955343 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwjrz\" (UniqueName: \"kubernetes.io/projected/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-kube-api-access-hwjrz\") on node \"crc\" DevicePath \"\"" Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.955386 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.955402 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.955415 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:06:40 crc kubenswrapper[4787]: I1203 18:06:40.955430 4787 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6c98e3-7b90-462f-ac9b-bf1874f97fc8-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.336862 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" event={"ID":"be6c98e3-7b90-462f-ac9b-bf1874f97fc8","Type":"ContainerDied","Data":"6e45bd11d3c380a8c96361741fcd3c0513a8f1ef2d26ef304a7e9cef23923db1"} Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.337198 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e45bd11d3c380a8c96361741fcd3c0513a8f1ef2d26ef304a7e9cef23923db1" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.336923 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.423912 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx"] Dec 03 18:06:41 crc kubenswrapper[4787]: E1203 18:06:41.424364 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be6c98e3-7b90-462f-ac9b-bf1874f97fc8" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.424383 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="be6c98e3-7b90-462f-ac9b-bf1874f97fc8" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.424597 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="be6c98e3-7b90-462f-ac9b-bf1874f97fc8" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.425276 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.428741 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.428765 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.428800 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.429077 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.429314 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.454877 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx"] Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.465183 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.465463 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.465598 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.465682 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ftxz\" (UniqueName: \"kubernetes.io/projected/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-kube-api-access-4ftxz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.465745 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.567629 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.567704 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.567746 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ftxz\" (UniqueName: \"kubernetes.io/projected/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-kube-api-access-4ftxz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.567781 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.567868 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.572414 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.572861 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.573136 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.574062 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.589189 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ftxz\" (UniqueName: \"kubernetes.io/projected/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-kube-api-access-4ftxz\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:41 crc kubenswrapper[4787]: I1203 18:06:41.741684 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:06:42 crc kubenswrapper[4787]: I1203 18:06:42.272339 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx"] Dec 03 18:06:42 crc kubenswrapper[4787]: I1203 18:06:42.348449 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" event={"ID":"9490b057-48b6-43c7-ae4f-3d09c1c3e16a","Type":"ContainerStarted","Data":"d48b42f01e34edd507840cafc7bda942bc1410e796961f815bfc45b0a0c14db2"} Dec 03 18:06:43 crc kubenswrapper[4787]: I1203 18:06:43.359501 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" event={"ID":"9490b057-48b6-43c7-ae4f-3d09c1c3e16a","Type":"ContainerStarted","Data":"e1277086ca07f5ed880117483fcfb2bb83864995c49506d004aa8f415cb1a2fc"} Dec 03 18:06:43 crc kubenswrapper[4787]: I1203 18:06:43.396218 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" podStartSLOduration=1.9613929049999999 podStartE2EDuration="2.396201091s" podCreationTimestamp="2025-12-03 18:06:41 +0000 UTC" firstStartedPulling="2025-12-03 18:06:42.277963532 +0000 UTC m=+3219.095434501" lastFinishedPulling="2025-12-03 18:06:42.712771688 +0000 UTC m=+3219.530242687" observedRunningTime="2025-12-03 18:06:43.39356202 +0000 UTC m=+3220.211032979" watchObservedRunningTime="2025-12-03 18:06:43.396201091 +0000 UTC m=+3220.213672050" Dec 03 18:06:51 crc kubenswrapper[4787]: I1203 18:06:51.766630 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:06:51 crc kubenswrapper[4787]: E1203 18:06:51.767301 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:07:03 crc kubenswrapper[4787]: I1203 18:07:03.775562 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:07:03 crc kubenswrapper[4787]: E1203 18:07:03.776346 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:07:18 crc kubenswrapper[4787]: I1203 18:07:18.766941 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:07:18 crc kubenswrapper[4787]: E1203 18:07:18.767768 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:07:26 crc kubenswrapper[4787]: I1203 18:07:26.592497 4787 scope.go:117] "RemoveContainer" containerID="dab7c1a7af8651abab7893e3de4bdf578dd9e6eb24ecaa656615c1042a292c03" Dec 03 18:07:30 crc kubenswrapper[4787]: I1203 18:07:30.767183 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:07:30 crc kubenswrapper[4787]: E1203 18:07:30.768223 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:07:41 crc kubenswrapper[4787]: I1203 18:07:41.771521 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:07:41 crc kubenswrapper[4787]: E1203 18:07:41.772476 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:07:55 crc kubenswrapper[4787]: I1203 18:07:55.765972 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:07:55 crc kubenswrapper[4787]: E1203 18:07:55.766783 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:08:10 crc kubenswrapper[4787]: I1203 18:08:10.767653 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:08:10 crc kubenswrapper[4787]: E1203 18:08:10.769005 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:08:23 crc kubenswrapper[4787]: I1203 18:08:23.776183 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:08:23 crc kubenswrapper[4787]: E1203 18:08:23.778214 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.664325 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lwdc7"] Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.667033 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.678394 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lwdc7"] Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.704247 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-catalog-content\") pod \"certified-operators-lwdc7\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.704316 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sf98v\" (UniqueName: \"kubernetes.io/projected/6b426b53-f6cb-409b-a8e0-848f5f647c03-kube-api-access-sf98v\") pod \"certified-operators-lwdc7\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.704399 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-utilities\") pod \"certified-operators-lwdc7\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.806088 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-catalog-content\") pod \"certified-operators-lwdc7\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.806167 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sf98v\" (UniqueName: \"kubernetes.io/projected/6b426b53-f6cb-409b-a8e0-848f5f647c03-kube-api-access-sf98v\") pod \"certified-operators-lwdc7\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.806249 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-utilities\") pod \"certified-operators-lwdc7\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.806630 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-catalog-content\") pod \"certified-operators-lwdc7\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.806673 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-utilities\") pod \"certified-operators-lwdc7\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.830218 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sf98v\" (UniqueName: \"kubernetes.io/projected/6b426b53-f6cb-409b-a8e0-848f5f647c03-kube-api-access-sf98v\") pod \"certified-operators-lwdc7\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:30 crc kubenswrapper[4787]: I1203 18:08:30.991735 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:31 crc kubenswrapper[4787]: I1203 18:08:31.557473 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lwdc7"] Dec 03 18:08:31 crc kubenswrapper[4787]: W1203 18:08:31.561267 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6b426b53_f6cb_409b_a8e0_848f5f647c03.slice/crio-90dd715726645c6eeb5de965c4bd6b40a48596ce94ac7b725b6af6806a11071d WatchSource:0}: Error finding container 90dd715726645c6eeb5de965c4bd6b40a48596ce94ac7b725b6af6806a11071d: Status 404 returned error can't find the container with id 90dd715726645c6eeb5de965c4bd6b40a48596ce94ac7b725b6af6806a11071d Dec 03 18:08:32 crc kubenswrapper[4787]: I1203 18:08:32.554743 4787 generic.go:334] "Generic (PLEG): container finished" podID="6b426b53-f6cb-409b-a8e0-848f5f647c03" containerID="b613fccc557371126d69543cab18a6b82e69ae5e86ed7796ef9dca8c858d2667" exitCode=0 Dec 03 18:08:32 crc kubenswrapper[4787]: I1203 18:08:32.554817 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwdc7" event={"ID":"6b426b53-f6cb-409b-a8e0-848f5f647c03","Type":"ContainerDied","Data":"b613fccc557371126d69543cab18a6b82e69ae5e86ed7796ef9dca8c858d2667"} Dec 03 18:08:32 crc kubenswrapper[4787]: I1203 18:08:32.555058 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwdc7" event={"ID":"6b426b53-f6cb-409b-a8e0-848f5f647c03","Type":"ContainerStarted","Data":"90dd715726645c6eeb5de965c4bd6b40a48596ce94ac7b725b6af6806a11071d"} Dec 03 18:08:33 crc kubenswrapper[4787]: I1203 18:08:33.576881 4787 generic.go:334] "Generic (PLEG): container finished" podID="6b426b53-f6cb-409b-a8e0-848f5f647c03" containerID="d5590838961ba5ae10f1ab35b012bb80cab510d4d0243713a204e33645eaef95" exitCode=0 Dec 03 18:08:33 crc kubenswrapper[4787]: I1203 18:08:33.577011 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwdc7" event={"ID":"6b426b53-f6cb-409b-a8e0-848f5f647c03","Type":"ContainerDied","Data":"d5590838961ba5ae10f1ab35b012bb80cab510d4d0243713a204e33645eaef95"} Dec 03 18:08:34 crc kubenswrapper[4787]: I1203 18:08:34.594909 4787 generic.go:334] "Generic (PLEG): container finished" podID="9490b057-48b6-43c7-ae4f-3d09c1c3e16a" containerID="e1277086ca07f5ed880117483fcfb2bb83864995c49506d004aa8f415cb1a2fc" exitCode=0 Dec 03 18:08:34 crc kubenswrapper[4787]: I1203 18:08:34.594972 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" event={"ID":"9490b057-48b6-43c7-ae4f-3d09c1c3e16a","Type":"ContainerDied","Data":"e1277086ca07f5ed880117483fcfb2bb83864995c49506d004aa8f415cb1a2fc"} Dec 03 18:08:34 crc kubenswrapper[4787]: I1203 18:08:34.599927 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwdc7" event={"ID":"6b426b53-f6cb-409b-a8e0-848f5f647c03","Type":"ContainerStarted","Data":"a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e"} Dec 03 18:08:34 crc kubenswrapper[4787]: I1203 18:08:34.644393 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lwdc7" podStartSLOduration=3.214917766 podStartE2EDuration="4.64437094s" podCreationTimestamp="2025-12-03 18:08:30 +0000 UTC" firstStartedPulling="2025-12-03 18:08:32.557012415 +0000 UTC m=+3329.374483374" lastFinishedPulling="2025-12-03 18:08:33.986465549 +0000 UTC m=+3330.803936548" observedRunningTime="2025-12-03 18:08:34.635092582 +0000 UTC m=+3331.452563551" watchObservedRunningTime="2025-12-03 18:08:34.64437094 +0000 UTC m=+3331.461841919" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.039887 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.138122 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ftxz\" (UniqueName: \"kubernetes.io/projected/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-kube-api-access-4ftxz\") pod \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.138252 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-bootstrap-combined-ca-bundle\") pod \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.138371 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ceph\") pod \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.138461 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-inventory\") pod \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.138482 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ssh-key\") pod \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\" (UID: \"9490b057-48b6-43c7-ae4f-3d09c1c3e16a\") " Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.143581 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "9490b057-48b6-43c7-ae4f-3d09c1c3e16a" (UID: "9490b057-48b6-43c7-ae4f-3d09c1c3e16a"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.151415 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ceph" (OuterVolumeSpecName: "ceph") pod "9490b057-48b6-43c7-ae4f-3d09c1c3e16a" (UID: "9490b057-48b6-43c7-ae4f-3d09c1c3e16a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.151442 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-kube-api-access-4ftxz" (OuterVolumeSpecName: "kube-api-access-4ftxz") pod "9490b057-48b6-43c7-ae4f-3d09c1c3e16a" (UID: "9490b057-48b6-43c7-ae4f-3d09c1c3e16a"). InnerVolumeSpecName "kube-api-access-4ftxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.173635 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-inventory" (OuterVolumeSpecName: "inventory") pod "9490b057-48b6-43c7-ae4f-3d09c1c3e16a" (UID: "9490b057-48b6-43c7-ae4f-3d09c1c3e16a"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.174167 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9490b057-48b6-43c7-ae4f-3d09c1c3e16a" (UID: "9490b057-48b6-43c7-ae4f-3d09c1c3e16a"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.240867 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.240899 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.240909 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ftxz\" (UniqueName: \"kubernetes.io/projected/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-kube-api-access-4ftxz\") on node \"crc\" DevicePath \"\"" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.240921 4787 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.240930 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9490b057-48b6-43c7-ae4f-3d09c1c3e16a-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.620968 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" event={"ID":"9490b057-48b6-43c7-ae4f-3d09c1c3e16a","Type":"ContainerDied","Data":"d48b42f01e34edd507840cafc7bda942bc1410e796961f815bfc45b0a0c14db2"} Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.621041 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d48b42f01e34edd507840cafc7bda942bc1410e796961f815bfc45b0a0c14db2" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.621073 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.720602 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj"] Dec 03 18:08:36 crc kubenswrapper[4787]: E1203 18:08:36.721394 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9490b057-48b6-43c7-ae4f-3d09c1c3e16a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.721428 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9490b057-48b6-43c7-ae4f-3d09c1c3e16a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.721725 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9490b057-48b6-43c7-ae4f-3d09c1c3e16a" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.723841 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.727451 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.727732 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.727817 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.727883 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.730591 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.740154 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj"] Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.767193 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:08:36 crc kubenswrapper[4787]: E1203 18:08:36.767823 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.854487 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.854655 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d25vb\" (UniqueName: \"kubernetes.io/projected/6df00b2f-8001-4386-83ed-830a938919e1-kube-api-access-d25vb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.854810 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.854847 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.957239 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.957342 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d25vb\" (UniqueName: \"kubernetes.io/projected/6df00b2f-8001-4386-83ed-830a938919e1-kube-api-access-d25vb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.957468 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.957517 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.961169 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.962210 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.962763 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:36 crc kubenswrapper[4787]: I1203 18:08:36.976160 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d25vb\" (UniqueName: \"kubernetes.io/projected/6df00b2f-8001-4386-83ed-830a938919e1-kube-api-access-d25vb\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:37 crc kubenswrapper[4787]: I1203 18:08:37.055110 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:08:37 crc kubenswrapper[4787]: I1203 18:08:37.620284 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj"] Dec 03 18:08:38 crc kubenswrapper[4787]: I1203 18:08:38.640296 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" event={"ID":"6df00b2f-8001-4386-83ed-830a938919e1","Type":"ContainerStarted","Data":"30c6a8e960f7600848f89422516d800eb8e5b6e70bc0156f57da30ee3b945650"} Dec 03 18:08:39 crc kubenswrapper[4787]: I1203 18:08:39.650839 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" event={"ID":"6df00b2f-8001-4386-83ed-830a938919e1","Type":"ContainerStarted","Data":"c5ca7431ff234d5afec499c846f98252ad8229bf8c26c6a8a3a1e751185c840e"} Dec 03 18:08:39 crc kubenswrapper[4787]: I1203 18:08:39.688735 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" podStartSLOduration=3.190246851 podStartE2EDuration="3.688711002s" podCreationTimestamp="2025-12-03 18:08:36 +0000 UTC" firstStartedPulling="2025-12-03 18:08:37.652069294 +0000 UTC m=+3334.469540253" lastFinishedPulling="2025-12-03 18:08:38.150533415 +0000 UTC m=+3334.968004404" observedRunningTime="2025-12-03 18:08:39.678819967 +0000 UTC m=+3336.496290916" watchObservedRunningTime="2025-12-03 18:08:39.688711002 +0000 UTC m=+3336.506181971" Dec 03 18:08:40 crc kubenswrapper[4787]: I1203 18:08:40.992515 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:40 crc kubenswrapper[4787]: I1203 18:08:40.993187 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:41 crc kubenswrapper[4787]: I1203 18:08:41.061691 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:41 crc kubenswrapper[4787]: I1203 18:08:41.725729 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:41 crc kubenswrapper[4787]: I1203 18:08:41.802214 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lwdc7"] Dec 03 18:08:43 crc kubenswrapper[4787]: I1203 18:08:43.689957 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lwdc7" podUID="6b426b53-f6cb-409b-a8e0-848f5f647c03" containerName="registry-server" containerID="cri-o://a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e" gracePeriod=2 Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.204580 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.316508 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-catalog-content\") pod \"6b426b53-f6cb-409b-a8e0-848f5f647c03\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.316834 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-utilities\") pod \"6b426b53-f6cb-409b-a8e0-848f5f647c03\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.316883 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sf98v\" (UniqueName: \"kubernetes.io/projected/6b426b53-f6cb-409b-a8e0-848f5f647c03-kube-api-access-sf98v\") pod \"6b426b53-f6cb-409b-a8e0-848f5f647c03\" (UID: \"6b426b53-f6cb-409b-a8e0-848f5f647c03\") " Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.317688 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-utilities" (OuterVolumeSpecName: "utilities") pod "6b426b53-f6cb-409b-a8e0-848f5f647c03" (UID: "6b426b53-f6cb-409b-a8e0-848f5f647c03"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.321576 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b426b53-f6cb-409b-a8e0-848f5f647c03-kube-api-access-sf98v" (OuterVolumeSpecName: "kube-api-access-sf98v") pod "6b426b53-f6cb-409b-a8e0-848f5f647c03" (UID: "6b426b53-f6cb-409b-a8e0-848f5f647c03"). InnerVolumeSpecName "kube-api-access-sf98v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.360393 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "6b426b53-f6cb-409b-a8e0-848f5f647c03" (UID: "6b426b53-f6cb-409b-a8e0-848f5f647c03"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.418972 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.419010 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6b426b53-f6cb-409b-a8e0-848f5f647c03-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.419038 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sf98v\" (UniqueName: \"kubernetes.io/projected/6b426b53-f6cb-409b-a8e0-848f5f647c03-kube-api-access-sf98v\") on node \"crc\" DevicePath \"\"" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.702392 4787 generic.go:334] "Generic (PLEG): container finished" podID="6b426b53-f6cb-409b-a8e0-848f5f647c03" containerID="a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e" exitCode=0 Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.702481 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lwdc7" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.702465 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwdc7" event={"ID":"6b426b53-f6cb-409b-a8e0-848f5f647c03","Type":"ContainerDied","Data":"a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e"} Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.702635 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lwdc7" event={"ID":"6b426b53-f6cb-409b-a8e0-848f5f647c03","Type":"ContainerDied","Data":"90dd715726645c6eeb5de965c4bd6b40a48596ce94ac7b725b6af6806a11071d"} Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.702682 4787 scope.go:117] "RemoveContainer" containerID="a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.724987 4787 scope.go:117] "RemoveContainer" containerID="d5590838961ba5ae10f1ab35b012bb80cab510d4d0243713a204e33645eaef95" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.744631 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lwdc7"] Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.761991 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lwdc7"] Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.769148 4787 scope.go:117] "RemoveContainer" containerID="b613fccc557371126d69543cab18a6b82e69ae5e86ed7796ef9dca8c858d2667" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.797875 4787 scope.go:117] "RemoveContainer" containerID="a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e" Dec 03 18:08:44 crc kubenswrapper[4787]: E1203 18:08:44.799111 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e\": container with ID starting with a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e not found: ID does not exist" containerID="a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.799165 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e"} err="failed to get container status \"a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e\": rpc error: code = NotFound desc = could not find container \"a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e\": container with ID starting with a4887578cace165abd016a36299ffdec45ea534e2388d5e62a74ab1ac230df9e not found: ID does not exist" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.799199 4787 scope.go:117] "RemoveContainer" containerID="d5590838961ba5ae10f1ab35b012bb80cab510d4d0243713a204e33645eaef95" Dec 03 18:08:44 crc kubenswrapper[4787]: E1203 18:08:44.799658 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5590838961ba5ae10f1ab35b012bb80cab510d4d0243713a204e33645eaef95\": container with ID starting with d5590838961ba5ae10f1ab35b012bb80cab510d4d0243713a204e33645eaef95 not found: ID does not exist" containerID="d5590838961ba5ae10f1ab35b012bb80cab510d4d0243713a204e33645eaef95" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.799713 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5590838961ba5ae10f1ab35b012bb80cab510d4d0243713a204e33645eaef95"} err="failed to get container status \"d5590838961ba5ae10f1ab35b012bb80cab510d4d0243713a204e33645eaef95\": rpc error: code = NotFound desc = could not find container \"d5590838961ba5ae10f1ab35b012bb80cab510d4d0243713a204e33645eaef95\": container with ID starting with d5590838961ba5ae10f1ab35b012bb80cab510d4d0243713a204e33645eaef95 not found: ID does not exist" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.799746 4787 scope.go:117] "RemoveContainer" containerID="b613fccc557371126d69543cab18a6b82e69ae5e86ed7796ef9dca8c858d2667" Dec 03 18:08:44 crc kubenswrapper[4787]: E1203 18:08:44.800132 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b613fccc557371126d69543cab18a6b82e69ae5e86ed7796ef9dca8c858d2667\": container with ID starting with b613fccc557371126d69543cab18a6b82e69ae5e86ed7796ef9dca8c858d2667 not found: ID does not exist" containerID="b613fccc557371126d69543cab18a6b82e69ae5e86ed7796ef9dca8c858d2667" Dec 03 18:08:44 crc kubenswrapper[4787]: I1203 18:08:44.800165 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b613fccc557371126d69543cab18a6b82e69ae5e86ed7796ef9dca8c858d2667"} err="failed to get container status \"b613fccc557371126d69543cab18a6b82e69ae5e86ed7796ef9dca8c858d2667\": rpc error: code = NotFound desc = could not find container \"b613fccc557371126d69543cab18a6b82e69ae5e86ed7796ef9dca8c858d2667\": container with ID starting with b613fccc557371126d69543cab18a6b82e69ae5e86ed7796ef9dca8c858d2667 not found: ID does not exist" Dec 03 18:08:45 crc kubenswrapper[4787]: I1203 18:08:45.787701 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b426b53-f6cb-409b-a8e0-848f5f647c03" path="/var/lib/kubelet/pods/6b426b53-f6cb-409b-a8e0-848f5f647c03/volumes" Dec 03 18:08:47 crc kubenswrapper[4787]: I1203 18:08:47.765843 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:08:47 crc kubenswrapper[4787]: E1203 18:08:47.766449 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:09:01 crc kubenswrapper[4787]: I1203 18:09:01.765875 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:09:01 crc kubenswrapper[4787]: E1203 18:09:01.766783 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:09:09 crc kubenswrapper[4787]: I1203 18:09:09.985531 4787 generic.go:334] "Generic (PLEG): container finished" podID="6df00b2f-8001-4386-83ed-830a938919e1" containerID="c5ca7431ff234d5afec499c846f98252ad8229bf8c26c6a8a3a1e751185c840e" exitCode=0 Dec 03 18:09:09 crc kubenswrapper[4787]: I1203 18:09:09.985600 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" event={"ID":"6df00b2f-8001-4386-83ed-830a938919e1","Type":"ContainerDied","Data":"c5ca7431ff234d5afec499c846f98252ad8229bf8c26c6a8a3a1e751185c840e"} Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.549577 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.632817 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ssh-key\") pod \"6df00b2f-8001-4386-83ed-830a938919e1\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.632856 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ceph\") pod \"6df00b2f-8001-4386-83ed-830a938919e1\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.632900 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d25vb\" (UniqueName: \"kubernetes.io/projected/6df00b2f-8001-4386-83ed-830a938919e1-kube-api-access-d25vb\") pod \"6df00b2f-8001-4386-83ed-830a938919e1\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.632923 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-inventory\") pod \"6df00b2f-8001-4386-83ed-830a938919e1\" (UID: \"6df00b2f-8001-4386-83ed-830a938919e1\") " Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.638488 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6df00b2f-8001-4386-83ed-830a938919e1-kube-api-access-d25vb" (OuterVolumeSpecName: "kube-api-access-d25vb") pod "6df00b2f-8001-4386-83ed-830a938919e1" (UID: "6df00b2f-8001-4386-83ed-830a938919e1"). InnerVolumeSpecName "kube-api-access-d25vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.638780 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ceph" (OuterVolumeSpecName: "ceph") pod "6df00b2f-8001-4386-83ed-830a938919e1" (UID: "6df00b2f-8001-4386-83ed-830a938919e1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.662893 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6df00b2f-8001-4386-83ed-830a938919e1" (UID: "6df00b2f-8001-4386-83ed-830a938919e1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.667425 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-inventory" (OuterVolumeSpecName: "inventory") pod "6df00b2f-8001-4386-83ed-830a938919e1" (UID: "6df00b2f-8001-4386-83ed-830a938919e1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.735839 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.735863 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.735872 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d25vb\" (UniqueName: \"kubernetes.io/projected/6df00b2f-8001-4386-83ed-830a938919e1-kube-api-access-d25vb\") on node \"crc\" DevicePath \"\"" Dec 03 18:09:11 crc kubenswrapper[4787]: I1203 18:09:11.735883 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6df00b2f-8001-4386-83ed-830a938919e1-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.009066 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" event={"ID":"6df00b2f-8001-4386-83ed-830a938919e1","Type":"ContainerDied","Data":"30c6a8e960f7600848f89422516d800eb8e5b6e70bc0156f57da30ee3b945650"} Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.009505 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="30c6a8e960f7600848f89422516d800eb8e5b6e70bc0156f57da30ee3b945650" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.009173 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.172180 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7"] Dec 03 18:09:12 crc kubenswrapper[4787]: E1203 18:09:12.172766 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b426b53-f6cb-409b-a8e0-848f5f647c03" containerName="extract-content" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.172846 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b426b53-f6cb-409b-a8e0-848f5f647c03" containerName="extract-content" Dec 03 18:09:12 crc kubenswrapper[4787]: E1203 18:09:12.172913 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6df00b2f-8001-4386-83ed-830a938919e1" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.173199 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6df00b2f-8001-4386-83ed-830a938919e1" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 18:09:12 crc kubenswrapper[4787]: E1203 18:09:12.173371 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b426b53-f6cb-409b-a8e0-848f5f647c03" containerName="extract-utilities" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.173444 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b426b53-f6cb-409b-a8e0-848f5f647c03" containerName="extract-utilities" Dec 03 18:09:12 crc kubenswrapper[4787]: E1203 18:09:12.173502 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b426b53-f6cb-409b-a8e0-848f5f647c03" containerName="registry-server" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.173549 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b426b53-f6cb-409b-a8e0-848f5f647c03" containerName="registry-server" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.173803 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6df00b2f-8001-4386-83ed-830a938919e1" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.173936 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b426b53-f6cb-409b-a8e0-848f5f647c03" containerName="registry-server" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.176902 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.181573 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.181776 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.181862 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.181964 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.182084 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.192396 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7"] Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.247401 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qkb92\" (UniqueName: \"kubernetes.io/projected/48e931e4-9962-4d61-ad24-62a2a5660510-kube-api-access-qkb92\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.247466 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.247550 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.247577 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.349092 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.349199 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.349222 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.349333 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qkb92\" (UniqueName: \"kubernetes.io/projected/48e931e4-9962-4d61-ad24-62a2a5660510-kube-api-access-qkb92\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.353730 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.353779 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.354149 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.365387 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qkb92\" (UniqueName: \"kubernetes.io/projected/48e931e4-9962-4d61-ad24-62a2a5660510-kube-api-access-qkb92\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:12 crc kubenswrapper[4787]: I1203 18:09:12.523427 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:13 crc kubenswrapper[4787]: I1203 18:09:13.146431 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7"] Dec 03 18:09:14 crc kubenswrapper[4787]: I1203 18:09:14.028714 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" event={"ID":"48e931e4-9962-4d61-ad24-62a2a5660510","Type":"ContainerStarted","Data":"7228d7bac14aa43a833bad0353d1456db187e7ba5bbb0d192e8a54f36a5371ce"} Dec 03 18:09:15 crc kubenswrapper[4787]: I1203 18:09:15.038854 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" event={"ID":"48e931e4-9962-4d61-ad24-62a2a5660510","Type":"ContainerStarted","Data":"985318343b2bfb84cde6dd128215977c26da5aad9556d7c3b3e8feaa6b03c6a6"} Dec 03 18:09:15 crc kubenswrapper[4787]: I1203 18:09:15.062300 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" podStartSLOduration=2.290635496 podStartE2EDuration="3.062281192s" podCreationTimestamp="2025-12-03 18:09:12 +0000 UTC" firstStartedPulling="2025-12-03 18:09:13.179211239 +0000 UTC m=+3369.996682198" lastFinishedPulling="2025-12-03 18:09:13.950856935 +0000 UTC m=+3370.768327894" observedRunningTime="2025-12-03 18:09:15.056142738 +0000 UTC m=+3371.873613697" watchObservedRunningTime="2025-12-03 18:09:15.062281192 +0000 UTC m=+3371.879752151" Dec 03 18:09:16 crc kubenswrapper[4787]: I1203 18:09:16.765839 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:09:16 crc kubenswrapper[4787]: E1203 18:09:16.766714 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:09:21 crc kubenswrapper[4787]: I1203 18:09:21.149340 4787 generic.go:334] "Generic (PLEG): container finished" podID="48e931e4-9962-4d61-ad24-62a2a5660510" containerID="985318343b2bfb84cde6dd128215977c26da5aad9556d7c3b3e8feaa6b03c6a6" exitCode=0 Dec 03 18:09:21 crc kubenswrapper[4787]: I1203 18:09:21.149483 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" event={"ID":"48e931e4-9962-4d61-ad24-62a2a5660510","Type":"ContainerDied","Data":"985318343b2bfb84cde6dd128215977c26da5aad9556d7c3b3e8feaa6b03c6a6"} Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.640605 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.789098 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-inventory\") pod \"48e931e4-9962-4d61-ad24-62a2a5660510\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.789165 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ceph\") pod \"48e931e4-9962-4d61-ad24-62a2a5660510\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.789195 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qkb92\" (UniqueName: \"kubernetes.io/projected/48e931e4-9962-4d61-ad24-62a2a5660510-kube-api-access-qkb92\") pod \"48e931e4-9962-4d61-ad24-62a2a5660510\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.789321 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ssh-key\") pod \"48e931e4-9962-4d61-ad24-62a2a5660510\" (UID: \"48e931e4-9962-4d61-ad24-62a2a5660510\") " Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.798009 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48e931e4-9962-4d61-ad24-62a2a5660510-kube-api-access-qkb92" (OuterVolumeSpecName: "kube-api-access-qkb92") pod "48e931e4-9962-4d61-ad24-62a2a5660510" (UID: "48e931e4-9962-4d61-ad24-62a2a5660510"). InnerVolumeSpecName "kube-api-access-qkb92". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.800075 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ceph" (OuterVolumeSpecName: "ceph") pod "48e931e4-9962-4d61-ad24-62a2a5660510" (UID: "48e931e4-9962-4d61-ad24-62a2a5660510"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.827301 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-inventory" (OuterVolumeSpecName: "inventory") pod "48e931e4-9962-4d61-ad24-62a2a5660510" (UID: "48e931e4-9962-4d61-ad24-62a2a5660510"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.829156 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "48e931e4-9962-4d61-ad24-62a2a5660510" (UID: "48e931e4-9962-4d61-ad24-62a2a5660510"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.892085 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.892131 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.892150 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qkb92\" (UniqueName: \"kubernetes.io/projected/48e931e4-9962-4d61-ad24-62a2a5660510-kube-api-access-qkb92\") on node \"crc\" DevicePath \"\"" Dec 03 18:09:22 crc kubenswrapper[4787]: I1203 18:09:22.892169 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/48e931e4-9962-4d61-ad24-62a2a5660510-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.173764 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" event={"ID":"48e931e4-9962-4d61-ad24-62a2a5660510","Type":"ContainerDied","Data":"7228d7bac14aa43a833bad0353d1456db187e7ba5bbb0d192e8a54f36a5371ce"} Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.174079 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7228d7bac14aa43a833bad0353d1456db187e7ba5bbb0d192e8a54f36a5371ce" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.173811 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.277916 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd"] Dec 03 18:09:23 crc kubenswrapper[4787]: E1203 18:09:23.278635 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48e931e4-9962-4d61-ad24-62a2a5660510" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.278666 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="48e931e4-9962-4d61-ad24-62a2a5660510" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.279089 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="48e931e4-9962-4d61-ad24-62a2a5660510" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.280237 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.284408 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.294598 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.294689 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.294729 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.296781 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.303054 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd"] Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.401291 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.401443 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.401617 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8tlzw\" (UniqueName: \"kubernetes.io/projected/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-kube-api-access-8tlzw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.401848 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.503493 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.503687 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.503778 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.503829 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8tlzw\" (UniqueName: \"kubernetes.io/projected/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-kube-api-access-8tlzw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.510493 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.510808 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.511512 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.525483 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8tlzw\" (UniqueName: \"kubernetes.io/projected/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-kube-api-access-8tlzw\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-6z8dd\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:23 crc kubenswrapper[4787]: I1203 18:09:23.598781 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:09:24 crc kubenswrapper[4787]: I1203 18:09:24.155775 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd"] Dec 03 18:09:24 crc kubenswrapper[4787]: I1203 18:09:24.184729 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" event={"ID":"5f10a9ab-33ee-4938-94f7-870cbdb92c9f","Type":"ContainerStarted","Data":"8bba4d16b235dc579d40a28bc9d8c9c04fe3b09ad02579f7b622d241682175cd"} Dec 03 18:09:25 crc kubenswrapper[4787]: I1203 18:09:25.195979 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" event={"ID":"5f10a9ab-33ee-4938-94f7-870cbdb92c9f","Type":"ContainerStarted","Data":"8e0a8084d0f27070509f8b07e6da2233b52361071b4fb71f0ef99c7284ec39e2"} Dec 03 18:09:30 crc kubenswrapper[4787]: I1203 18:09:30.767189 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:09:30 crc kubenswrapper[4787]: E1203 18:09:30.768447 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:09:44 crc kubenswrapper[4787]: I1203 18:09:44.766948 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:09:44 crc kubenswrapper[4787]: E1203 18:09:44.767773 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:09:55 crc kubenswrapper[4787]: I1203 18:09:55.766507 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:09:55 crc kubenswrapper[4787]: E1203 18:09:55.767429 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:10:07 crc kubenswrapper[4787]: I1203 18:10:07.766207 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:10:07 crc kubenswrapper[4787]: E1203 18:10:07.767215 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:10:10 crc kubenswrapper[4787]: I1203 18:10:10.697267 4787 generic.go:334] "Generic (PLEG): container finished" podID="5f10a9ab-33ee-4938-94f7-870cbdb92c9f" containerID="8e0a8084d0f27070509f8b07e6da2233b52361071b4fb71f0ef99c7284ec39e2" exitCode=0 Dec 03 18:10:10 crc kubenswrapper[4787]: I1203 18:10:10.697453 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" event={"ID":"5f10a9ab-33ee-4938-94f7-870cbdb92c9f","Type":"ContainerDied","Data":"8e0a8084d0f27070509f8b07e6da2233b52361071b4fb71f0ef99c7284ec39e2"} Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.195946 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.299584 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tlzw\" (UniqueName: \"kubernetes.io/projected/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-kube-api-access-8tlzw\") pod \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.299715 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ceph\") pod \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.299794 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-inventory\") pod \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.299823 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ssh-key\") pod \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\" (UID: \"5f10a9ab-33ee-4938-94f7-870cbdb92c9f\") " Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.305727 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ceph" (OuterVolumeSpecName: "ceph") pod "5f10a9ab-33ee-4938-94f7-870cbdb92c9f" (UID: "5f10a9ab-33ee-4938-94f7-870cbdb92c9f"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.307099 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-kube-api-access-8tlzw" (OuterVolumeSpecName: "kube-api-access-8tlzw") pod "5f10a9ab-33ee-4938-94f7-870cbdb92c9f" (UID: "5f10a9ab-33ee-4938-94f7-870cbdb92c9f"). InnerVolumeSpecName "kube-api-access-8tlzw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.338696 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-inventory" (OuterVolumeSpecName: "inventory") pod "5f10a9ab-33ee-4938-94f7-870cbdb92c9f" (UID: "5f10a9ab-33ee-4938-94f7-870cbdb92c9f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.340950 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "5f10a9ab-33ee-4938-94f7-870cbdb92c9f" (UID: "5f10a9ab-33ee-4938-94f7-870cbdb92c9f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.402937 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tlzw\" (UniqueName: \"kubernetes.io/projected/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-kube-api-access-8tlzw\") on node \"crc\" DevicePath \"\"" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.403245 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.403260 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.403272 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/5f10a9ab-33ee-4938-94f7-870cbdb92c9f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.727524 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" event={"ID":"5f10a9ab-33ee-4938-94f7-870cbdb92c9f","Type":"ContainerDied","Data":"8bba4d16b235dc579d40a28bc9d8c9c04fe3b09ad02579f7b622d241682175cd"} Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.727736 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bba4d16b235dc579d40a28bc9d8c9c04fe3b09ad02579f7b622d241682175cd" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.727621 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-6z8dd" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.824065 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw"] Dec 03 18:10:12 crc kubenswrapper[4787]: E1203 18:10:12.824595 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5f10a9ab-33ee-4938-94f7-870cbdb92c9f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.824620 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5f10a9ab-33ee-4938-94f7-870cbdb92c9f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.824910 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5f10a9ab-33ee-4938-94f7-870cbdb92c9f" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.825914 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.829237 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.829396 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.829696 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.829895 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.829959 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.842976 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw"] Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.912911 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.913192 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hvd4m\" (UniqueName: \"kubernetes.io/projected/9034ed5d-53f4-4db9-9421-516a5945fce9-kube-api-access-hvd4m\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.913238 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:12 crc kubenswrapper[4787]: I1203 18:10:12.913270 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:13 crc kubenswrapper[4787]: I1203 18:10:13.015031 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hvd4m\" (UniqueName: \"kubernetes.io/projected/9034ed5d-53f4-4db9-9421-516a5945fce9-kube-api-access-hvd4m\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:13 crc kubenswrapper[4787]: I1203 18:10:13.015097 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:13 crc kubenswrapper[4787]: I1203 18:10:13.015128 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:13 crc kubenswrapper[4787]: I1203 18:10:13.015162 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:13 crc kubenswrapper[4787]: I1203 18:10:13.019381 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:13 crc kubenswrapper[4787]: I1203 18:10:13.020274 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:13 crc kubenswrapper[4787]: I1203 18:10:13.028095 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:13 crc kubenswrapper[4787]: I1203 18:10:13.031579 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hvd4m\" (UniqueName: \"kubernetes.io/projected/9034ed5d-53f4-4db9-9421-516a5945fce9-kube-api-access-hvd4m\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:13 crc kubenswrapper[4787]: I1203 18:10:13.163374 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:13 crc kubenswrapper[4787]: I1203 18:10:13.751995 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw"] Dec 03 18:10:14 crc kubenswrapper[4787]: I1203 18:10:14.749709 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" event={"ID":"9034ed5d-53f4-4db9-9421-516a5945fce9","Type":"ContainerStarted","Data":"8a8fe293ca100995972097f6dafce9dc2202e59a339a1116d293d39a527acd34"} Dec 03 18:10:14 crc kubenswrapper[4787]: I1203 18:10:14.750535 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" event={"ID":"9034ed5d-53f4-4db9-9421-516a5945fce9","Type":"ContainerStarted","Data":"d7ca7ce1c7c012dd8b65ac70217f1c03f4fef091059f13a323ffbec832cd1afb"} Dec 03 18:10:14 crc kubenswrapper[4787]: I1203 18:10:14.767260 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" podStartSLOduration=2.372371407 podStartE2EDuration="2.767242303s" podCreationTimestamp="2025-12-03 18:10:12 +0000 UTC" firstStartedPulling="2025-12-03 18:10:13.756896354 +0000 UTC m=+3430.574367333" lastFinishedPulling="2025-12-03 18:10:14.15176726 +0000 UTC m=+3430.969238229" observedRunningTime="2025-12-03 18:10:14.765510447 +0000 UTC m=+3431.582981416" watchObservedRunningTime="2025-12-03 18:10:14.767242303 +0000 UTC m=+3431.584713272" Dec 03 18:10:18 crc kubenswrapper[4787]: I1203 18:10:18.765791 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:10:18 crc kubenswrapper[4787]: E1203 18:10:18.766739 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.047467 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-km8t2"] Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.063257 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.065875 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-km8t2"] Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.165522 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8cbz6\" (UniqueName: \"kubernetes.io/projected/18ae3c33-33de-4260-86f5-254e0f097e46-kube-api-access-8cbz6\") pod \"redhat-marketplace-km8t2\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.166581 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-catalog-content\") pod \"redhat-marketplace-km8t2\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.166637 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-utilities\") pod \"redhat-marketplace-km8t2\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.268012 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-catalog-content\") pod \"redhat-marketplace-km8t2\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.268362 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-utilities\") pod \"redhat-marketplace-km8t2\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.268510 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8cbz6\" (UniqueName: \"kubernetes.io/projected/18ae3c33-33de-4260-86f5-254e0f097e46-kube-api-access-8cbz6\") pod \"redhat-marketplace-km8t2\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.268750 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-catalog-content\") pod \"redhat-marketplace-km8t2\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.269054 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-utilities\") pod \"redhat-marketplace-km8t2\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.288768 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8cbz6\" (UniqueName: \"kubernetes.io/projected/18ae3c33-33de-4260-86f5-254e0f097e46-kube-api-access-8cbz6\") pod \"redhat-marketplace-km8t2\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.407473 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.801320 4787 generic.go:334] "Generic (PLEG): container finished" podID="9034ed5d-53f4-4db9-9421-516a5945fce9" containerID="8a8fe293ca100995972097f6dafce9dc2202e59a339a1116d293d39a527acd34" exitCode=0 Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.801424 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" event={"ID":"9034ed5d-53f4-4db9-9421-516a5945fce9","Type":"ContainerDied","Data":"8a8fe293ca100995972097f6dafce9dc2202e59a339a1116d293d39a527acd34"} Dec 03 18:10:19 crc kubenswrapper[4787]: I1203 18:10:19.893352 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-km8t2"] Dec 03 18:10:20 crc kubenswrapper[4787]: I1203 18:10:20.817290 4787 generic.go:334] "Generic (PLEG): container finished" podID="18ae3c33-33de-4260-86f5-254e0f097e46" containerID="2d71ffa6f2a62a95a245fa476ea1ad8d52081393ab5855154c4eb33224a6c85b" exitCode=0 Dec 03 18:10:20 crc kubenswrapper[4787]: I1203 18:10:20.817419 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km8t2" event={"ID":"18ae3c33-33de-4260-86f5-254e0f097e46","Type":"ContainerDied","Data":"2d71ffa6f2a62a95a245fa476ea1ad8d52081393ab5855154c4eb33224a6c85b"} Dec 03 18:10:20 crc kubenswrapper[4787]: I1203 18:10:20.817684 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km8t2" event={"ID":"18ae3c33-33de-4260-86f5-254e0f097e46","Type":"ContainerStarted","Data":"f5e773496f60f9d9479e2e0908b260274d52ab72a913b2ab2a70ac43093bc1b8"} Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.285155 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.410700 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ssh-key\") pod \"9034ed5d-53f4-4db9-9421-516a5945fce9\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.410911 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ceph\") pod \"9034ed5d-53f4-4db9-9421-516a5945fce9\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.411064 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hvd4m\" (UniqueName: \"kubernetes.io/projected/9034ed5d-53f4-4db9-9421-516a5945fce9-kube-api-access-hvd4m\") pod \"9034ed5d-53f4-4db9-9421-516a5945fce9\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.411106 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-inventory\") pod \"9034ed5d-53f4-4db9-9421-516a5945fce9\" (UID: \"9034ed5d-53f4-4db9-9421-516a5945fce9\") " Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.418159 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9034ed5d-53f4-4db9-9421-516a5945fce9-kube-api-access-hvd4m" (OuterVolumeSpecName: "kube-api-access-hvd4m") pod "9034ed5d-53f4-4db9-9421-516a5945fce9" (UID: "9034ed5d-53f4-4db9-9421-516a5945fce9"). InnerVolumeSpecName "kube-api-access-hvd4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.418944 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ceph" (OuterVolumeSpecName: "ceph") pod "9034ed5d-53f4-4db9-9421-516a5945fce9" (UID: "9034ed5d-53f4-4db9-9421-516a5945fce9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.449888 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9034ed5d-53f4-4db9-9421-516a5945fce9" (UID: "9034ed5d-53f4-4db9-9421-516a5945fce9"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.457076 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-inventory" (OuterVolumeSpecName: "inventory") pod "9034ed5d-53f4-4db9-9421-516a5945fce9" (UID: "9034ed5d-53f4-4db9-9421-516a5945fce9"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.513573 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hvd4m\" (UniqueName: \"kubernetes.io/projected/9034ed5d-53f4-4db9-9421-516a5945fce9-kube-api-access-hvd4m\") on node \"crc\" DevicePath \"\"" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.513623 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.513640 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.513656 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9034ed5d-53f4-4db9-9421-516a5945fce9-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.828787 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" event={"ID":"9034ed5d-53f4-4db9-9421-516a5945fce9","Type":"ContainerDied","Data":"d7ca7ce1c7c012dd8b65ac70217f1c03f4fef091059f13a323ffbec832cd1afb"} Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.829095 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d7ca7ce1c7c012dd8b65ac70217f1c03f4fef091059f13a323ffbec832cd1afb" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.828810 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.833166 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km8t2" event={"ID":"18ae3c33-33de-4260-86f5-254e0f097e46","Type":"ContainerStarted","Data":"27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8"} Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.907480 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8"] Dec 03 18:10:21 crc kubenswrapper[4787]: E1203 18:10:21.907952 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9034ed5d-53f4-4db9-9421-516a5945fce9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.907970 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9034ed5d-53f4-4db9-9421-516a5945fce9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.908220 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9034ed5d-53f4-4db9-9421-516a5945fce9" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.909003 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.911185 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.910959 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.911627 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.911897 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.919659 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:10:21 crc kubenswrapper[4787]: I1203 18:10:21.932494 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8"] Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.024218 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lwmvx\" (UniqueName: \"kubernetes.io/projected/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-kube-api-access-lwmvx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.024993 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.025362 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.025424 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.127787 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.127954 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.128145 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.128352 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lwmvx\" (UniqueName: \"kubernetes.io/projected/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-kube-api-access-lwmvx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.135775 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.135816 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.143380 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.156092 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lwmvx\" (UniqueName: \"kubernetes.io/projected/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-kube-api-access-lwmvx\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.224220 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.796732 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8"] Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.843927 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" event={"ID":"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6","Type":"ContainerStarted","Data":"99374134d32df6356568b6b633bb1ec29b33dbfd8cf2d01872f61c5284953bac"} Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.847251 4787 generic.go:334] "Generic (PLEG): container finished" podID="18ae3c33-33de-4260-86f5-254e0f097e46" containerID="27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8" exitCode=0 Dec 03 18:10:22 crc kubenswrapper[4787]: I1203 18:10:22.847288 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km8t2" event={"ID":"18ae3c33-33de-4260-86f5-254e0f097e46","Type":"ContainerDied","Data":"27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8"} Dec 03 18:10:23 crc kubenswrapper[4787]: I1203 18:10:23.856948 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" event={"ID":"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6","Type":"ContainerStarted","Data":"c33fb80aeac86c38c37c55835d8010867d45bbe36e439c08562a9d4711755da9"} Dec 03 18:10:23 crc kubenswrapper[4787]: I1203 18:10:23.860714 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km8t2" event={"ID":"18ae3c33-33de-4260-86f5-254e0f097e46","Type":"ContainerStarted","Data":"221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a"} Dec 03 18:10:23 crc kubenswrapper[4787]: I1203 18:10:23.880527 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" podStartSLOduration=2.480387096 podStartE2EDuration="2.880508953s" podCreationTimestamp="2025-12-03 18:10:21 +0000 UTC" firstStartedPulling="2025-12-03 18:10:22.803214079 +0000 UTC m=+3439.620685058" lastFinishedPulling="2025-12-03 18:10:23.203335946 +0000 UTC m=+3440.020806915" observedRunningTime="2025-12-03 18:10:23.876934767 +0000 UTC m=+3440.694405746" watchObservedRunningTime="2025-12-03 18:10:23.880508953 +0000 UTC m=+3440.697979912" Dec 03 18:10:23 crc kubenswrapper[4787]: I1203 18:10:23.898518 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-km8t2" podStartSLOduration=2.465189975 podStartE2EDuration="4.898503565s" podCreationTimestamp="2025-12-03 18:10:19 +0000 UTC" firstStartedPulling="2025-12-03 18:10:20.821374681 +0000 UTC m=+3437.638845670" lastFinishedPulling="2025-12-03 18:10:23.254688301 +0000 UTC m=+3440.072159260" observedRunningTime="2025-12-03 18:10:23.893365957 +0000 UTC m=+3440.710836916" watchObservedRunningTime="2025-12-03 18:10:23.898503565 +0000 UTC m=+3440.715974524" Dec 03 18:10:29 crc kubenswrapper[4787]: I1203 18:10:29.407598 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:29 crc kubenswrapper[4787]: I1203 18:10:29.408282 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:29 crc kubenswrapper[4787]: I1203 18:10:29.484906 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:29 crc kubenswrapper[4787]: I1203 18:10:29.970927 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:30 crc kubenswrapper[4787]: I1203 18:10:30.017950 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-km8t2"] Dec 03 18:10:31 crc kubenswrapper[4787]: I1203 18:10:31.934877 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-km8t2" podUID="18ae3c33-33de-4260-86f5-254e0f097e46" containerName="registry-server" containerID="cri-o://221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a" gracePeriod=2 Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.527191 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.667528 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-utilities\") pod \"18ae3c33-33de-4260-86f5-254e0f097e46\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.667575 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-catalog-content\") pod \"18ae3c33-33de-4260-86f5-254e0f097e46\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.667673 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8cbz6\" (UniqueName: \"kubernetes.io/projected/18ae3c33-33de-4260-86f5-254e0f097e46-kube-api-access-8cbz6\") pod \"18ae3c33-33de-4260-86f5-254e0f097e46\" (UID: \"18ae3c33-33de-4260-86f5-254e0f097e46\") " Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.668861 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-utilities" (OuterVolumeSpecName: "utilities") pod "18ae3c33-33de-4260-86f5-254e0f097e46" (UID: "18ae3c33-33de-4260-86f5-254e0f097e46"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.674294 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18ae3c33-33de-4260-86f5-254e0f097e46-kube-api-access-8cbz6" (OuterVolumeSpecName: "kube-api-access-8cbz6") pod "18ae3c33-33de-4260-86f5-254e0f097e46" (UID: "18ae3c33-33de-4260-86f5-254e0f097e46"). InnerVolumeSpecName "kube-api-access-8cbz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.690804 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "18ae3c33-33de-4260-86f5-254e0f097e46" (UID: "18ae3c33-33de-4260-86f5-254e0f097e46"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.765793 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.771395 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.771441 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8cbz6\" (UniqueName: \"kubernetes.io/projected/18ae3c33-33de-4260-86f5-254e0f097e46-kube-api-access-8cbz6\") on node \"crc\" DevicePath \"\"" Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.771455 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/18ae3c33-33de-4260-86f5-254e0f097e46-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.947389 4787 generic.go:334] "Generic (PLEG): container finished" podID="18ae3c33-33de-4260-86f5-254e0f097e46" containerID="221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a" exitCode=0 Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.947438 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km8t2" event={"ID":"18ae3c33-33de-4260-86f5-254e0f097e46","Type":"ContainerDied","Data":"221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a"} Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.947475 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-km8t2" event={"ID":"18ae3c33-33de-4260-86f5-254e0f097e46","Type":"ContainerDied","Data":"f5e773496f60f9d9479e2e0908b260274d52ab72a913b2ab2a70ac43093bc1b8"} Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.947495 4787 scope.go:117] "RemoveContainer" containerID="221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a" Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.947544 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-km8t2" Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.974744 4787 scope.go:117] "RemoveContainer" containerID="27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8" Dec 03 18:10:32 crc kubenswrapper[4787]: I1203 18:10:32.989249 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-km8t2"] Dec 03 18:10:33 crc kubenswrapper[4787]: I1203 18:10:32.998511 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-km8t2"] Dec 03 18:10:33 crc kubenswrapper[4787]: I1203 18:10:33.043933 4787 scope.go:117] "RemoveContainer" containerID="2d71ffa6f2a62a95a245fa476ea1ad8d52081393ab5855154c4eb33224a6c85b" Dec 03 18:10:33 crc kubenswrapper[4787]: I1203 18:10:33.131091 4787 scope.go:117] "RemoveContainer" containerID="221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a" Dec 03 18:10:33 crc kubenswrapper[4787]: E1203 18:10:33.146433 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a\": container with ID starting with 221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a not found: ID does not exist" containerID="221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a" Dec 03 18:10:33 crc kubenswrapper[4787]: I1203 18:10:33.146482 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a"} err="failed to get container status \"221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a\": rpc error: code = NotFound desc = could not find container \"221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a\": container with ID starting with 221271a5fe79a546ef7beb210125b28432f971cb7aa701f41aa783f73ecaba9a not found: ID does not exist" Dec 03 18:10:33 crc kubenswrapper[4787]: I1203 18:10:33.146508 4787 scope.go:117] "RemoveContainer" containerID="27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8" Dec 03 18:10:33 crc kubenswrapper[4787]: E1203 18:10:33.155490 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8\": container with ID starting with 27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8 not found: ID does not exist" containerID="27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8" Dec 03 18:10:33 crc kubenswrapper[4787]: I1203 18:10:33.155559 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8"} err="failed to get container status \"27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8\": rpc error: code = NotFound desc = could not find container \"27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8\": container with ID starting with 27119d5fccb2f53f87fc560d6849839cd15e53ec708717e711dae70458fce2a8 not found: ID does not exist" Dec 03 18:10:33 crc kubenswrapper[4787]: I1203 18:10:33.155592 4787 scope.go:117] "RemoveContainer" containerID="2d71ffa6f2a62a95a245fa476ea1ad8d52081393ab5855154c4eb33224a6c85b" Dec 03 18:10:33 crc kubenswrapper[4787]: E1203 18:10:33.167381 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2d71ffa6f2a62a95a245fa476ea1ad8d52081393ab5855154c4eb33224a6c85b\": container with ID starting with 2d71ffa6f2a62a95a245fa476ea1ad8d52081393ab5855154c4eb33224a6c85b not found: ID does not exist" containerID="2d71ffa6f2a62a95a245fa476ea1ad8d52081393ab5855154c4eb33224a6c85b" Dec 03 18:10:33 crc kubenswrapper[4787]: I1203 18:10:33.167449 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2d71ffa6f2a62a95a245fa476ea1ad8d52081393ab5855154c4eb33224a6c85b"} err="failed to get container status \"2d71ffa6f2a62a95a245fa476ea1ad8d52081393ab5855154c4eb33224a6c85b\": rpc error: code = NotFound desc = could not find container \"2d71ffa6f2a62a95a245fa476ea1ad8d52081393ab5855154c4eb33224a6c85b\": container with ID starting with 2d71ffa6f2a62a95a245fa476ea1ad8d52081393ab5855154c4eb33224a6c85b not found: ID does not exist" Dec 03 18:10:33 crc kubenswrapper[4787]: I1203 18:10:33.785303 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18ae3c33-33de-4260-86f5-254e0f097e46" path="/var/lib/kubelet/pods/18ae3c33-33de-4260-86f5-254e0f097e46/volumes" Dec 03 18:10:33 crc kubenswrapper[4787]: I1203 18:10:33.960680 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"8027146e8048a2924f6ea539e853b86d156ce857b489636ed54a2f77398dc20e"} Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.514334 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x8ppz"] Dec 03 18:11:18 crc kubenswrapper[4787]: E1203 18:11:18.525937 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ae3c33-33de-4260-86f5-254e0f097e46" containerName="registry-server" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.525960 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ae3c33-33de-4260-86f5-254e0f097e46" containerName="registry-server" Dec 03 18:11:18 crc kubenswrapper[4787]: E1203 18:11:18.526028 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ae3c33-33de-4260-86f5-254e0f097e46" containerName="extract-utilities" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.526038 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ae3c33-33de-4260-86f5-254e0f097e46" containerName="extract-utilities" Dec 03 18:11:18 crc kubenswrapper[4787]: E1203 18:11:18.526068 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="18ae3c33-33de-4260-86f5-254e0f097e46" containerName="extract-content" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.526076 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="18ae3c33-33de-4260-86f5-254e0f097e46" containerName="extract-content" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.526351 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="18ae3c33-33de-4260-86f5-254e0f097e46" containerName="registry-server" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.528290 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x8ppz"] Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.558696 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.692323 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-catalog-content\") pod \"redhat-operators-x8ppz\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.692465 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f86f5\" (UniqueName: \"kubernetes.io/projected/8425bd45-79f7-4a09-bc8a-810d8ff675fd-kube-api-access-f86f5\") pod \"redhat-operators-x8ppz\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.692539 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-utilities\") pod \"redhat-operators-x8ppz\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.795099 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f86f5\" (UniqueName: \"kubernetes.io/projected/8425bd45-79f7-4a09-bc8a-810d8ff675fd-kube-api-access-f86f5\") pod \"redhat-operators-x8ppz\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.795265 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-utilities\") pod \"redhat-operators-x8ppz\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.795333 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-catalog-content\") pod \"redhat-operators-x8ppz\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.796307 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-catalog-content\") pod \"redhat-operators-x8ppz\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.797113 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-utilities\") pod \"redhat-operators-x8ppz\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.841859 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f86f5\" (UniqueName: \"kubernetes.io/projected/8425bd45-79f7-4a09-bc8a-810d8ff675fd-kube-api-access-f86f5\") pod \"redhat-operators-x8ppz\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:18 crc kubenswrapper[4787]: I1203 18:11:18.927683 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:19 crc kubenswrapper[4787]: W1203 18:11:19.550914 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8425bd45_79f7_4a09_bc8a_810d8ff675fd.slice/crio-bae1ab2ba610254aac71b5e696c1aff427fd6dcaf75dfb86540bb3d7c75ede9b WatchSource:0}: Error finding container bae1ab2ba610254aac71b5e696c1aff427fd6dcaf75dfb86540bb3d7c75ede9b: Status 404 returned error can't find the container with id bae1ab2ba610254aac71b5e696c1aff427fd6dcaf75dfb86540bb3d7c75ede9b Dec 03 18:11:19 crc kubenswrapper[4787]: I1203 18:11:19.560329 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x8ppz"] Dec 03 18:11:20 crc kubenswrapper[4787]: I1203 18:11:20.559464 4787 generic.go:334] "Generic (PLEG): container finished" podID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerID="7a76356f4243410e1a69d1ca5811433dde51d76e2d91a6fb87dfbae286499623" exitCode=0 Dec 03 18:11:20 crc kubenswrapper[4787]: I1203 18:11:20.559515 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8ppz" event={"ID":"8425bd45-79f7-4a09-bc8a-810d8ff675fd","Type":"ContainerDied","Data":"7a76356f4243410e1a69d1ca5811433dde51d76e2d91a6fb87dfbae286499623"} Dec 03 18:11:20 crc kubenswrapper[4787]: I1203 18:11:20.559545 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8ppz" event={"ID":"8425bd45-79f7-4a09-bc8a-810d8ff675fd","Type":"ContainerStarted","Data":"bae1ab2ba610254aac71b5e696c1aff427fd6dcaf75dfb86540bb3d7c75ede9b"} Dec 03 18:11:21 crc kubenswrapper[4787]: I1203 18:11:21.571503 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8ppz" event={"ID":"8425bd45-79f7-4a09-bc8a-810d8ff675fd","Type":"ContainerStarted","Data":"be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70"} Dec 03 18:11:23 crc kubenswrapper[4787]: I1203 18:11:23.594802 4787 generic.go:334] "Generic (PLEG): container finished" podID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerID="be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70" exitCode=0 Dec 03 18:11:23 crc kubenswrapper[4787]: I1203 18:11:23.594876 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8ppz" event={"ID":"8425bd45-79f7-4a09-bc8a-810d8ff675fd","Type":"ContainerDied","Data":"be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70"} Dec 03 18:11:23 crc kubenswrapper[4787]: I1203 18:11:23.597286 4787 generic.go:334] "Generic (PLEG): container finished" podID="d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6" containerID="c33fb80aeac86c38c37c55835d8010867d45bbe36e439c08562a9d4711755da9" exitCode=0 Dec 03 18:11:23 crc kubenswrapper[4787]: I1203 18:11:23.597328 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" event={"ID":"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6","Type":"ContainerDied","Data":"c33fb80aeac86c38c37c55835d8010867d45bbe36e439c08562a9d4711755da9"} Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.178087 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.233719 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ceph\") pod \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.233871 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-inventory\") pod \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.233902 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ssh-key\") pod \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.233951 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lwmvx\" (UniqueName: \"kubernetes.io/projected/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-kube-api-access-lwmvx\") pod \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\" (UID: \"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6\") " Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.239576 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-kube-api-access-lwmvx" (OuterVolumeSpecName: "kube-api-access-lwmvx") pod "d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6" (UID: "d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6"). InnerVolumeSpecName "kube-api-access-lwmvx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.239616 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ceph" (OuterVolumeSpecName: "ceph") pod "d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6" (UID: "d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.264355 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6" (UID: "d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.270136 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-inventory" (OuterVolumeSpecName: "inventory") pod "d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6" (UID: "d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.338193 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.338795 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.338812 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.338828 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lwmvx\" (UniqueName: \"kubernetes.io/projected/d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6-kube-api-access-lwmvx\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.624661 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.624661 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8" event={"ID":"d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6","Type":"ContainerDied","Data":"99374134d32df6356568b6b633bb1ec29b33dbfd8cf2d01872f61c5284953bac"} Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.624815 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="99374134d32df6356568b6b633bb1ec29b33dbfd8cf2d01872f61c5284953bac" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.628364 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8ppz" event={"ID":"8425bd45-79f7-4a09-bc8a-810d8ff675fd","Type":"ContainerStarted","Data":"887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2"} Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.661832 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x8ppz" podStartSLOduration=4.359283257 podStartE2EDuration="8.66181419s" podCreationTimestamp="2025-12-03 18:11:17 +0000 UTC" firstStartedPulling="2025-12-03 18:11:20.561126124 +0000 UTC m=+3497.378597083" lastFinishedPulling="2025-12-03 18:11:24.863657057 +0000 UTC m=+3501.681128016" observedRunningTime="2025-12-03 18:11:25.655368087 +0000 UTC m=+3502.472839056" watchObservedRunningTime="2025-12-03 18:11:25.66181419 +0000 UTC m=+3502.479285159" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.743584 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-g87t4"] Dec 03 18:11:25 crc kubenswrapper[4787]: E1203 18:11:25.744123 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.744145 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.744409 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.745287 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.747717 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.748279 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.748288 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.748303 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.748737 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.763649 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-g87t4"] Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.848651 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.848753 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.848897 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j6zwv\" (UniqueName: \"kubernetes.io/projected/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-kube-api-access-j6zwv\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.849230 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ceph\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.951328 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.951466 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.951593 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j6zwv\" (UniqueName: \"kubernetes.io/projected/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-kube-api-access-j6zwv\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.951769 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ceph\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.956929 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.957080 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.970896 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ceph\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:25 crc kubenswrapper[4787]: I1203 18:11:25.971330 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j6zwv\" (UniqueName: \"kubernetes.io/projected/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-kube-api-access-j6zwv\") pod \"ssh-known-hosts-edpm-deployment-g87t4\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:26 crc kubenswrapper[4787]: I1203 18:11:26.061551 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:26 crc kubenswrapper[4787]: I1203 18:11:26.732207 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-g87t4"] Dec 03 18:11:26 crc kubenswrapper[4787]: I1203 18:11:26.736404 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:11:27 crc kubenswrapper[4787]: I1203 18:11:27.656128 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" event={"ID":"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac","Type":"ContainerStarted","Data":"e7567922071aab893131878eb6b044eff414113bc97caf5e3e13220ea023df68"} Dec 03 18:11:27 crc kubenswrapper[4787]: I1203 18:11:27.656715 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" event={"ID":"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac","Type":"ContainerStarted","Data":"6ad1de54f1989afd7e53f6cc04ebc1db5c882dbf00e9e32339553e2886855741"} Dec 03 18:11:27 crc kubenswrapper[4787]: I1203 18:11:27.682316 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" podStartSLOduration=2.23966283 podStartE2EDuration="2.682294933s" podCreationTimestamp="2025-12-03 18:11:25 +0000 UTC" firstStartedPulling="2025-12-03 18:11:26.736088435 +0000 UTC m=+3503.553559404" lastFinishedPulling="2025-12-03 18:11:27.178720548 +0000 UTC m=+3503.996191507" observedRunningTime="2025-12-03 18:11:27.676386295 +0000 UTC m=+3504.493857264" watchObservedRunningTime="2025-12-03 18:11:27.682294933 +0000 UTC m=+3504.499765892" Dec 03 18:11:28 crc kubenswrapper[4787]: I1203 18:11:28.927863 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:28 crc kubenswrapper[4787]: I1203 18:11:28.927944 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:29 crc kubenswrapper[4787]: I1203 18:11:29.985231 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x8ppz" podUID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerName="registry-server" probeResult="failure" output=< Dec 03 18:11:29 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 18:11:29 crc kubenswrapper[4787]: > Dec 03 18:11:38 crc kubenswrapper[4787]: I1203 18:11:38.803307 4787 generic.go:334] "Generic (PLEG): container finished" podID="9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac" containerID="e7567922071aab893131878eb6b044eff414113bc97caf5e3e13220ea023df68" exitCode=0 Dec 03 18:11:38 crc kubenswrapper[4787]: I1203 18:11:38.803599 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" event={"ID":"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac","Type":"ContainerDied","Data":"e7567922071aab893131878eb6b044eff414113bc97caf5e3e13220ea023df68"} Dec 03 18:11:38 crc kubenswrapper[4787]: I1203 18:11:38.977987 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:39 crc kubenswrapper[4787]: I1203 18:11:39.026692 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:39 crc kubenswrapper[4787]: I1203 18:11:39.211172 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x8ppz"] Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.248753 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.377378 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ssh-key-openstack-edpm-ipam\") pod \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.377438 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ceph\") pod \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.377532 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-inventory-0\") pod \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.378372 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j6zwv\" (UniqueName: \"kubernetes.io/projected/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-kube-api-access-j6zwv\") pod \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\" (UID: \"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac\") " Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.383493 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ceph" (OuterVolumeSpecName: "ceph") pod "9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac" (UID: "9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.383832 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-kube-api-access-j6zwv" (OuterVolumeSpecName: "kube-api-access-j6zwv") pod "9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac" (UID: "9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac"). InnerVolumeSpecName "kube-api-access-j6zwv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.410565 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac" (UID: "9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.410839 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac" (UID: "9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.480754 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.480784 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.480796 4787 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.480807 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j6zwv\" (UniqueName: \"kubernetes.io/projected/9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac-kube-api-access-j6zwv\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.824893 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.824952 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-g87t4" event={"ID":"9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac","Type":"ContainerDied","Data":"6ad1de54f1989afd7e53f6cc04ebc1db5c882dbf00e9e32339553e2886855741"} Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.824979 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6ad1de54f1989afd7e53f6cc04ebc1db5c882dbf00e9e32339553e2886855741" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.825105 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x8ppz" podUID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerName="registry-server" containerID="cri-o://887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2" gracePeriod=2 Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.918914 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d"] Dec 03 18:11:40 crc kubenswrapper[4787]: E1203 18:11:40.919366 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac" containerName="ssh-known-hosts-edpm-deployment" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.919386 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac" containerName="ssh-known-hosts-edpm-deployment" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.919632 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac" containerName="ssh-known-hosts-edpm-deployment" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.920350 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.922587 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.923148 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.923345 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.923391 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.923275 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.931066 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d"] Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.989541 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2llq\" (UniqueName: \"kubernetes.io/projected/8fbd7464-0bdd-4280-87a4-f4b14b82b057-kube-api-access-x2llq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.989936 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.990012 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:40 crc kubenswrapper[4787]: I1203 18:11:40.990081 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.092231 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2llq\" (UniqueName: \"kubernetes.io/projected/8fbd7464-0bdd-4280-87a4-f4b14b82b057-kube-api-access-x2llq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.097684 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.097756 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.097822 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.104695 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.105237 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.105514 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.108321 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2llq\" (UniqueName: \"kubernetes.io/projected/8fbd7464-0bdd-4280-87a4-f4b14b82b057-kube-api-access-x2llq\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-pfw4d\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.275668 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.303070 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.409078 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-utilities\") pod \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.409622 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f86f5\" (UniqueName: \"kubernetes.io/projected/8425bd45-79f7-4a09-bc8a-810d8ff675fd-kube-api-access-f86f5\") pod \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.409860 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-catalog-content\") pod \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\" (UID: \"8425bd45-79f7-4a09-bc8a-810d8ff675fd\") " Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.410215 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-utilities" (OuterVolumeSpecName: "utilities") pod "8425bd45-79f7-4a09-bc8a-810d8ff675fd" (UID: "8425bd45-79f7-4a09-bc8a-810d8ff675fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.410657 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.417843 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8425bd45-79f7-4a09-bc8a-810d8ff675fd-kube-api-access-f86f5" (OuterVolumeSpecName: "kube-api-access-f86f5") pod "8425bd45-79f7-4a09-bc8a-810d8ff675fd" (UID: "8425bd45-79f7-4a09-bc8a-810d8ff675fd"). InnerVolumeSpecName "kube-api-access-f86f5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.512668 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f86f5\" (UniqueName: \"kubernetes.io/projected/8425bd45-79f7-4a09-bc8a-810d8ff675fd-kube-api-access-f86f5\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.521869 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8425bd45-79f7-4a09-bc8a-810d8ff675fd" (UID: "8425bd45-79f7-4a09-bc8a-810d8ff675fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.616666 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8425bd45-79f7-4a09-bc8a-810d8ff675fd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.838320 4787 generic.go:334] "Generic (PLEG): container finished" podID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerID="887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2" exitCode=0 Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.838412 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8ppz" event={"ID":"8425bd45-79f7-4a09-bc8a-810d8ff675fd","Type":"ContainerDied","Data":"887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2"} Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.838496 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x8ppz" event={"ID":"8425bd45-79f7-4a09-bc8a-810d8ff675fd","Type":"ContainerDied","Data":"bae1ab2ba610254aac71b5e696c1aff427fd6dcaf75dfb86540bb3d7c75ede9b"} Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.838527 4787 scope.go:117] "RemoveContainer" containerID="887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.838756 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x8ppz" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.869359 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d"] Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.872525 4787 scope.go:117] "RemoveContainer" containerID="be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.879713 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x8ppz"] Dec 03 18:11:41 crc kubenswrapper[4787]: W1203 18:11:41.884816 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8fbd7464_0bdd_4280_87a4_f4b14b82b057.slice/crio-298ab62e823a995d18f631e893bc879ea7a35e842c8b783e4a549994b9a3d211 WatchSource:0}: Error finding container 298ab62e823a995d18f631e893bc879ea7a35e842c8b783e4a549994b9a3d211: Status 404 returned error can't find the container with id 298ab62e823a995d18f631e893bc879ea7a35e842c8b783e4a549994b9a3d211 Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.891898 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x8ppz"] Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.898992 4787 scope.go:117] "RemoveContainer" containerID="7a76356f4243410e1a69d1ca5811433dde51d76e2d91a6fb87dfbae286499623" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.930796 4787 scope.go:117] "RemoveContainer" containerID="887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2" Dec 03 18:11:41 crc kubenswrapper[4787]: E1203 18:11:41.931357 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2\": container with ID starting with 887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2 not found: ID does not exist" containerID="887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.931402 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2"} err="failed to get container status \"887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2\": rpc error: code = NotFound desc = could not find container \"887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2\": container with ID starting with 887f3126b54d0350a7d3518c8c3f4db2df979ae3534b95d3ff3fbe94b9a69cd2 not found: ID does not exist" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.931430 4787 scope.go:117] "RemoveContainer" containerID="be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70" Dec 03 18:11:41 crc kubenswrapper[4787]: E1203 18:11:41.931919 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70\": container with ID starting with be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70 not found: ID does not exist" containerID="be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.931949 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70"} err="failed to get container status \"be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70\": rpc error: code = NotFound desc = could not find container \"be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70\": container with ID starting with be625431aba27b7fe626a66f1710bd75903fecfe7e4152710373b6dcfb55bf70 not found: ID does not exist" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.931963 4787 scope.go:117] "RemoveContainer" containerID="7a76356f4243410e1a69d1ca5811433dde51d76e2d91a6fb87dfbae286499623" Dec 03 18:11:41 crc kubenswrapper[4787]: E1203 18:11:41.932361 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a76356f4243410e1a69d1ca5811433dde51d76e2d91a6fb87dfbae286499623\": container with ID starting with 7a76356f4243410e1a69d1ca5811433dde51d76e2d91a6fb87dfbae286499623 not found: ID does not exist" containerID="7a76356f4243410e1a69d1ca5811433dde51d76e2d91a6fb87dfbae286499623" Dec 03 18:11:41 crc kubenswrapper[4787]: I1203 18:11:41.932402 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a76356f4243410e1a69d1ca5811433dde51d76e2d91a6fb87dfbae286499623"} err="failed to get container status \"7a76356f4243410e1a69d1ca5811433dde51d76e2d91a6fb87dfbae286499623\": rpc error: code = NotFound desc = could not find container \"7a76356f4243410e1a69d1ca5811433dde51d76e2d91a6fb87dfbae286499623\": container with ID starting with 7a76356f4243410e1a69d1ca5811433dde51d76e2d91a6fb87dfbae286499623 not found: ID does not exist" Dec 03 18:11:42 crc kubenswrapper[4787]: I1203 18:11:42.858165 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" event={"ID":"8fbd7464-0bdd-4280-87a4-f4b14b82b057","Type":"ContainerStarted","Data":"dee1ec01cb6d7e4c6ab59362b319a4f84106732545384317dcb516e172f3c39f"} Dec 03 18:11:42 crc kubenswrapper[4787]: I1203 18:11:42.858887 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" event={"ID":"8fbd7464-0bdd-4280-87a4-f4b14b82b057","Type":"ContainerStarted","Data":"298ab62e823a995d18f631e893bc879ea7a35e842c8b783e4a549994b9a3d211"} Dec 03 18:11:42 crc kubenswrapper[4787]: I1203 18:11:42.888066 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" podStartSLOduration=2.469935121 podStartE2EDuration="2.888047617s" podCreationTimestamp="2025-12-03 18:11:40 +0000 UTC" firstStartedPulling="2025-12-03 18:11:41.898994732 +0000 UTC m=+3518.716465691" lastFinishedPulling="2025-12-03 18:11:42.317107228 +0000 UTC m=+3519.134578187" observedRunningTime="2025-12-03 18:11:42.87284536 +0000 UTC m=+3519.690316359" watchObservedRunningTime="2025-12-03 18:11:42.888047617 +0000 UTC m=+3519.705518576" Dec 03 18:11:43 crc kubenswrapper[4787]: I1203 18:11:43.788085 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" path="/var/lib/kubelet/pods/8425bd45-79f7-4a09-bc8a-810d8ff675fd/volumes" Dec 03 18:11:51 crc kubenswrapper[4787]: I1203 18:11:51.979707 4787 generic.go:334] "Generic (PLEG): container finished" podID="8fbd7464-0bdd-4280-87a4-f4b14b82b057" containerID="dee1ec01cb6d7e4c6ab59362b319a4f84106732545384317dcb516e172f3c39f" exitCode=0 Dec 03 18:11:51 crc kubenswrapper[4787]: I1203 18:11:51.979777 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" event={"ID":"8fbd7464-0bdd-4280-87a4-f4b14b82b057","Type":"ContainerDied","Data":"dee1ec01cb6d7e4c6ab59362b319a4f84106732545384317dcb516e172f3c39f"} Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.398787 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.489279 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2llq\" (UniqueName: \"kubernetes.io/projected/8fbd7464-0bdd-4280-87a4-f4b14b82b057-kube-api-access-x2llq\") pod \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.489380 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ssh-key\") pod \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.489421 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-inventory\") pod \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.489438 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ceph\") pod \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\" (UID: \"8fbd7464-0bdd-4280-87a4-f4b14b82b057\") " Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.495968 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ceph" (OuterVolumeSpecName: "ceph") pod "8fbd7464-0bdd-4280-87a4-f4b14b82b057" (UID: "8fbd7464-0bdd-4280-87a4-f4b14b82b057"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.496392 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8fbd7464-0bdd-4280-87a4-f4b14b82b057-kube-api-access-x2llq" (OuterVolumeSpecName: "kube-api-access-x2llq") pod "8fbd7464-0bdd-4280-87a4-f4b14b82b057" (UID: "8fbd7464-0bdd-4280-87a4-f4b14b82b057"). InnerVolumeSpecName "kube-api-access-x2llq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.526145 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8fbd7464-0bdd-4280-87a4-f4b14b82b057" (UID: "8fbd7464-0bdd-4280-87a4-f4b14b82b057"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.526756 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-inventory" (OuterVolumeSpecName: "inventory") pod "8fbd7464-0bdd-4280-87a4-f4b14b82b057" (UID: "8fbd7464-0bdd-4280-87a4-f4b14b82b057"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.591452 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2llq\" (UniqueName: \"kubernetes.io/projected/8fbd7464-0bdd-4280-87a4-f4b14b82b057-kube-api-access-x2llq\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.591492 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.591506 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:53 crc kubenswrapper[4787]: I1203 18:11:53.591518 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8fbd7464-0bdd-4280-87a4-f4b14b82b057-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.005690 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" event={"ID":"8fbd7464-0bdd-4280-87a4-f4b14b82b057","Type":"ContainerDied","Data":"298ab62e823a995d18f631e893bc879ea7a35e842c8b783e4a549994b9a3d211"} Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.006060 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="298ab62e823a995d18f631e893bc879ea7a35e842c8b783e4a549994b9a3d211" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.005807 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-pfw4d" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.100514 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6"] Dec 03 18:11:54 crc kubenswrapper[4787]: E1203 18:11:54.100977 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerName="registry-server" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.100993 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerName="registry-server" Dec 03 18:11:54 crc kubenswrapper[4787]: E1203 18:11:54.101043 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerName="extract-content" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.101051 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerName="extract-content" Dec 03 18:11:54 crc kubenswrapper[4787]: E1203 18:11:54.101061 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerName="extract-utilities" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.101070 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerName="extract-utilities" Dec 03 18:11:54 crc kubenswrapper[4787]: E1203 18:11:54.101090 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8fbd7464-0bdd-4280-87a4-f4b14b82b057" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.101099 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8fbd7464-0bdd-4280-87a4-f4b14b82b057" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.101291 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8fbd7464-0bdd-4280-87a4-f4b14b82b057" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.101308 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8425bd45-79f7-4a09-bc8a-810d8ff675fd" containerName="registry-server" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.102083 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.104149 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.104553 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.104647 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.104762 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.104825 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.114080 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6"] Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.205721 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqgxq\" (UniqueName: \"kubernetes.io/projected/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-kube-api-access-pqgxq\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.205849 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.205907 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.206048 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.308323 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqgxq\" (UniqueName: \"kubernetes.io/projected/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-kube-api-access-pqgxq\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.308458 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.308501 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.309254 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.312897 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.312929 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.313646 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.323640 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqgxq\" (UniqueName: \"kubernetes.io/projected/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-kube-api-access-pqgxq\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.434941 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:11:54 crc kubenswrapper[4787]: I1203 18:11:54.946684 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6"] Dec 03 18:11:55 crc kubenswrapper[4787]: I1203 18:11:55.017676 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" event={"ID":"ce4c2dba-e3f9-4745-adbe-92d44c7bb025","Type":"ContainerStarted","Data":"1f90b65d770598321a7f8f019d30bb117408b768824a385091871698c40fcb40"} Dec 03 18:11:56 crc kubenswrapper[4787]: I1203 18:11:56.032700 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" event={"ID":"ce4c2dba-e3f9-4745-adbe-92d44c7bb025","Type":"ContainerStarted","Data":"2f6a95438524ffda88dd6623e4bfb466ed3b5983aa7302c1fe099d5a08c17af8"} Dec 03 18:11:56 crc kubenswrapper[4787]: I1203 18:11:56.058933 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" podStartSLOduration=1.640053803 podStartE2EDuration="2.058916529s" podCreationTimestamp="2025-12-03 18:11:54 +0000 UTC" firstStartedPulling="2025-12-03 18:11:54.957056884 +0000 UTC m=+3531.774527843" lastFinishedPulling="2025-12-03 18:11:55.37591961 +0000 UTC m=+3532.193390569" observedRunningTime="2025-12-03 18:11:56.05594965 +0000 UTC m=+3532.873420649" watchObservedRunningTime="2025-12-03 18:11:56.058916529 +0000 UTC m=+3532.876387488" Dec 03 18:12:07 crc kubenswrapper[4787]: I1203 18:12:07.159776 4787 generic.go:334] "Generic (PLEG): container finished" podID="ce4c2dba-e3f9-4745-adbe-92d44c7bb025" containerID="2f6a95438524ffda88dd6623e4bfb466ed3b5983aa7302c1fe099d5a08c17af8" exitCode=0 Dec 03 18:12:07 crc kubenswrapper[4787]: I1203 18:12:07.159877 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" event={"ID":"ce4c2dba-e3f9-4745-adbe-92d44c7bb025","Type":"ContainerDied","Data":"2f6a95438524ffda88dd6623e4bfb466ed3b5983aa7302c1fe099d5a08c17af8"} Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.623996 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.719584 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-inventory\") pod \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.720096 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pqgxq\" (UniqueName: \"kubernetes.io/projected/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-kube-api-access-pqgxq\") pod \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.720370 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ceph\") pod \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.720642 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ssh-key\") pod \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\" (UID: \"ce4c2dba-e3f9-4745-adbe-92d44c7bb025\") " Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.726173 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ceph" (OuterVolumeSpecName: "ceph") pod "ce4c2dba-e3f9-4745-adbe-92d44c7bb025" (UID: "ce4c2dba-e3f9-4745-adbe-92d44c7bb025"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.726416 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-kube-api-access-pqgxq" (OuterVolumeSpecName: "kube-api-access-pqgxq") pod "ce4c2dba-e3f9-4745-adbe-92d44c7bb025" (UID: "ce4c2dba-e3f9-4745-adbe-92d44c7bb025"). InnerVolumeSpecName "kube-api-access-pqgxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.750799 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ce4c2dba-e3f9-4745-adbe-92d44c7bb025" (UID: "ce4c2dba-e3f9-4745-adbe-92d44c7bb025"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.756289 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-inventory" (OuterVolumeSpecName: "inventory") pod "ce4c2dba-e3f9-4745-adbe-92d44c7bb025" (UID: "ce4c2dba-e3f9-4745-adbe-92d44c7bb025"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.823740 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.823813 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.823840 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pqgxq\" (UniqueName: \"kubernetes.io/projected/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-kube-api-access-pqgxq\") on node \"crc\" DevicePath \"\"" Dec 03 18:12:08 crc kubenswrapper[4787]: I1203 18:12:08.824087 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ce4c2dba-e3f9-4745-adbe-92d44c7bb025-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.184841 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" event={"ID":"ce4c2dba-e3f9-4745-adbe-92d44c7bb025","Type":"ContainerDied","Data":"1f90b65d770598321a7f8f019d30bb117408b768824a385091871698c40fcb40"} Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.184903 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1f90b65d770598321a7f8f019d30bb117408b768824a385091871698c40fcb40" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.184930 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.296469 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7"] Dec 03 18:12:09 crc kubenswrapper[4787]: E1203 18:12:09.297204 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce4c2dba-e3f9-4745-adbe-92d44c7bb025" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.297225 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce4c2dba-e3f9-4745-adbe-92d44c7bb025" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.297457 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce4c2dba-e3f9-4745-adbe-92d44c7bb025" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.298147 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.300646 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.301147 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.302334 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.303370 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.303670 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.304158 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.304459 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.304694 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.305448 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.312219 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7"] Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.312318 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-telemetry-default-certs-0" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.438979 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439071 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439112 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439159 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439205 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439315 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439344 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439368 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439404 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439441 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439511 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439543 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439592 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439632 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jvv2\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-kube-api-access-4jvv2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439723 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439771 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.439802 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.541908 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.541990 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.542040 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.542082 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.542117 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.542148 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.542190 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.542760 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.542815 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.542855 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.542885 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.542925 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.542967 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.543039 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.543073 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.543159 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.547827 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.548479 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.548706 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.549890 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-power-monitoring-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.550922 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.550710 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.551355 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.552253 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.553392 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jvv2\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-kube-api-access-4jvv2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.553557 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.556109 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.556295 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.556558 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.557340 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.557444 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.557456 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.563433 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.584591 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jvv2\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-kube-api-access-4jvv2\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:09 crc kubenswrapper[4787]: I1203 18:12:09.623379 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:12:10 crc kubenswrapper[4787]: I1203 18:12:10.227597 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7"] Dec 03 18:12:10 crc kubenswrapper[4787]: W1203 18:12:10.235121 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6d97cc3a_8056_4f1a_98c8_a9dedb9a905d.slice/crio-df09ca3c2138442d5579fc84d8950d404cfd8d5fbb2b331de7bbb84349073bf6 WatchSource:0}: Error finding container df09ca3c2138442d5579fc84d8950d404cfd8d5fbb2b331de7bbb84349073bf6: Status 404 returned error can't find the container with id df09ca3c2138442d5579fc84d8950d404cfd8d5fbb2b331de7bbb84349073bf6 Dec 03 18:12:11 crc kubenswrapper[4787]: I1203 18:12:11.222350 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" event={"ID":"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d","Type":"ContainerStarted","Data":"af19549ea316f496a5e9e883876224008fa6573780c9249dac8ac3d7db39a261"} Dec 03 18:12:11 crc kubenswrapper[4787]: I1203 18:12:11.223060 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" event={"ID":"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d","Type":"ContainerStarted","Data":"df09ca3c2138442d5579fc84d8950d404cfd8d5fbb2b331de7bbb84349073bf6"} Dec 03 18:12:11 crc kubenswrapper[4787]: I1203 18:12:11.257304 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" podStartSLOduration=1.843642689 podStartE2EDuration="2.257281896s" podCreationTimestamp="2025-12-03 18:12:09 +0000 UTC" firstStartedPulling="2025-12-03 18:12:10.238585037 +0000 UTC m=+3547.056055986" lastFinishedPulling="2025-12-03 18:12:10.652224194 +0000 UTC m=+3547.469695193" observedRunningTime="2025-12-03 18:12:11.251381698 +0000 UTC m=+3548.068852667" watchObservedRunningTime="2025-12-03 18:12:11.257281896 +0000 UTC m=+3548.074752865" Dec 03 18:12:48 crc kubenswrapper[4787]: I1203 18:12:48.989395 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:12:48 crc kubenswrapper[4787]: I1203 18:12:48.990995 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:13:15 crc kubenswrapper[4787]: I1203 18:13:15.019812 4787 generic.go:334] "Generic (PLEG): container finished" podID="6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" containerID="af19549ea316f496a5e9e883876224008fa6573780c9249dac8ac3d7db39a261" exitCode=0 Dec 03 18:13:15 crc kubenswrapper[4787]: I1203 18:13:15.019890 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" event={"ID":"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d","Type":"ContainerDied","Data":"af19549ea316f496a5e9e883876224008fa6573780c9249dac8ac3d7db39a261"} Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.516171 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.609628 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-power-monitoring-combined-ca-bundle\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.609728 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-nova-combined-ca-bundle\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.609769 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4jvv2\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-kube-api-access-4jvv2\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.609853 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-neutron-metadata-combined-ca-bundle\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.609893 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-libvirt-combined-ca-bundle\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.609935 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-ovn-default-certs-0\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.609989 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.610078 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-bootstrap-combined-ca-bundle\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.610162 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.610214 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ceph\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.610308 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ssh-key\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.610345 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.610391 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ovn-combined-ca-bundle\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.610452 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-combined-ca-bundle\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.610581 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-default-certs-0\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.610620 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-repo-setup-combined-ca-bundle\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.610649 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-inventory\") pod \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\" (UID: \"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d\") " Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.617961 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.618076 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.618160 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.619104 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.619823 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-telemetry-default-certs-0") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "openstack-edpm-ipam-telemetry-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.619917 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.620818 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.622711 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.622883 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.623235 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-kube-api-access-4jvv2" (OuterVolumeSpecName: "kube-api-access-4jvv2") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "kube-api-access-4jvv2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.623782 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.624982 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.626200 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ceph" (OuterVolumeSpecName: "ceph") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.626335 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.628408 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.654927 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-inventory" (OuterVolumeSpecName: "inventory") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.658803 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" (UID: "6d97cc3a-8056-4f1a-98c8-a9dedb9a905d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.713558 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.713797 4787 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.713886 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.714166 4787 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.714471 4787 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.714608 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4jvv2\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-kube-api-access-4jvv2\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.714703 4787 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.714781 4787 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.714856 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.714944 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.715051 4787 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.715170 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-telemetry-power-monitoring-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.715281 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.715360 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.715446 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.715525 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:16 crc kubenswrapper[4787]: I1203 18:13:16.715610 4787 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6d97cc3a-8056-4f1a-98c8-a9dedb9a905d-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.047067 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" event={"ID":"6d97cc3a-8056-4f1a-98c8-a9dedb9a905d","Type":"ContainerDied","Data":"df09ca3c2138442d5579fc84d8950d404cfd8d5fbb2b331de7bbb84349073bf6"} Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.047104 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.047115 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df09ca3c2138442d5579fc84d8950d404cfd8d5fbb2b331de7bbb84349073bf6" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.208418 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr"] Dec 03 18:13:17 crc kubenswrapper[4787]: E1203 18:13:17.209561 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.209606 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.210090 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d97cc3a-8056-4f1a-98c8-a9dedb9a905d" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.211399 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.214867 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.215144 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.215244 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.215396 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.220275 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.232809 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr"] Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.330921 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.331092 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.331148 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.331286 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hf2kp\" (UniqueName: \"kubernetes.io/projected/862d8a45-30d2-480c-9aa2-2b8019465b79-kube-api-access-hf2kp\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.433059 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.433174 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.433212 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.433342 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hf2kp\" (UniqueName: \"kubernetes.io/projected/862d8a45-30d2-480c-9aa2-2b8019465b79-kube-api-access-hf2kp\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.437969 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.438604 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.439726 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.451252 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hf2kp\" (UniqueName: \"kubernetes.io/projected/862d8a45-30d2-480c-9aa2-2b8019465b79-kube-api-access-hf2kp\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:17 crc kubenswrapper[4787]: I1203 18:13:17.537954 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:18 crc kubenswrapper[4787]: I1203 18:13:18.109987 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr"] Dec 03 18:13:18 crc kubenswrapper[4787]: I1203 18:13:18.989580 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:13:18 crc kubenswrapper[4787]: I1203 18:13:18.989874 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:13:19 crc kubenswrapper[4787]: I1203 18:13:19.071102 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" event={"ID":"862d8a45-30d2-480c-9aa2-2b8019465b79","Type":"ContainerStarted","Data":"c02382bb8b7160e4aac193b1a0fc659643a7bb68f1d065403461b60f148aea3a"} Dec 03 18:13:19 crc kubenswrapper[4787]: I1203 18:13:19.071156 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" event={"ID":"862d8a45-30d2-480c-9aa2-2b8019465b79","Type":"ContainerStarted","Data":"e7e24aaf73c403f59f0921c9a27aea469c345bb9153f7b37737abff08d6a2d0a"} Dec 03 18:13:19 crc kubenswrapper[4787]: I1203 18:13:19.097514 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" podStartSLOduration=1.504595373 podStartE2EDuration="2.097493159s" podCreationTimestamp="2025-12-03 18:13:17 +0000 UTC" firstStartedPulling="2025-12-03 18:13:18.120949889 +0000 UTC m=+3614.938420888" lastFinishedPulling="2025-12-03 18:13:18.713847675 +0000 UTC m=+3615.531318674" observedRunningTime="2025-12-03 18:13:19.088706763 +0000 UTC m=+3615.906177722" watchObservedRunningTime="2025-12-03 18:13:19.097493159 +0000 UTC m=+3615.914964118" Dec 03 18:13:26 crc kubenswrapper[4787]: I1203 18:13:26.167796 4787 generic.go:334] "Generic (PLEG): container finished" podID="862d8a45-30d2-480c-9aa2-2b8019465b79" containerID="c02382bb8b7160e4aac193b1a0fc659643a7bb68f1d065403461b60f148aea3a" exitCode=0 Dec 03 18:13:26 crc kubenswrapper[4787]: I1203 18:13:26.167943 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" event={"ID":"862d8a45-30d2-480c-9aa2-2b8019465b79","Type":"ContainerDied","Data":"c02382bb8b7160e4aac193b1a0fc659643a7bb68f1d065403461b60f148aea3a"} Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.689873 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.786906 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ssh-key\") pod \"862d8a45-30d2-480c-9aa2-2b8019465b79\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.786990 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-inventory\") pod \"862d8a45-30d2-480c-9aa2-2b8019465b79\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.787045 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hf2kp\" (UniqueName: \"kubernetes.io/projected/862d8a45-30d2-480c-9aa2-2b8019465b79-kube-api-access-hf2kp\") pod \"862d8a45-30d2-480c-9aa2-2b8019465b79\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.787099 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ceph\") pod \"862d8a45-30d2-480c-9aa2-2b8019465b79\" (UID: \"862d8a45-30d2-480c-9aa2-2b8019465b79\") " Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.792718 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/862d8a45-30d2-480c-9aa2-2b8019465b79-kube-api-access-hf2kp" (OuterVolumeSpecName: "kube-api-access-hf2kp") pod "862d8a45-30d2-480c-9aa2-2b8019465b79" (UID: "862d8a45-30d2-480c-9aa2-2b8019465b79"). InnerVolumeSpecName "kube-api-access-hf2kp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.792765 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ceph" (OuterVolumeSpecName: "ceph") pod "862d8a45-30d2-480c-9aa2-2b8019465b79" (UID: "862d8a45-30d2-480c-9aa2-2b8019465b79"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.816238 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-inventory" (OuterVolumeSpecName: "inventory") pod "862d8a45-30d2-480c-9aa2-2b8019465b79" (UID: "862d8a45-30d2-480c-9aa2-2b8019465b79"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.827763 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "862d8a45-30d2-480c-9aa2-2b8019465b79" (UID: "862d8a45-30d2-480c-9aa2-2b8019465b79"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.890295 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.890330 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.890344 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hf2kp\" (UniqueName: \"kubernetes.io/projected/862d8a45-30d2-480c-9aa2-2b8019465b79-kube-api-access-hf2kp\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:27 crc kubenswrapper[4787]: I1203 18:13:27.890353 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/862d8a45-30d2-480c-9aa2-2b8019465b79-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.208475 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" event={"ID":"862d8a45-30d2-480c-9aa2-2b8019465b79","Type":"ContainerDied","Data":"e7e24aaf73c403f59f0921c9a27aea469c345bb9153f7b37737abff08d6a2d0a"} Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.208548 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e7e24aaf73c403f59f0921c9a27aea469c345bb9153f7b37737abff08d6a2d0a" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.208621 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.277672 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt"] Dec 03 18:13:28 crc kubenswrapper[4787]: E1203 18:13:28.278354 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862d8a45-30d2-480c-9aa2-2b8019465b79" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.278386 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="862d8a45-30d2-480c-9aa2-2b8019465b79" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.278734 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="862d8a45-30d2-480c-9aa2-2b8019465b79" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.279907 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.292775 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.292853 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.292779 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.293075 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.293121 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.294904 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.302917 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt"] Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.408047 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.408360 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.408391 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.408435 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.408457 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lq95t\" (UniqueName: \"kubernetes.io/projected/4fe91a7c-36b7-4528-8061-f2c47bede56d-kube-api-access-lq95t\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.408629 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.510860 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.510928 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.510997 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.511046 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lq95t\" (UniqueName: \"kubernetes.io/projected/4fe91a7c-36b7-4528-8061-f2c47bede56d-kube-api-access-lq95t\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.511091 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.511227 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.512910 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.514863 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.514908 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.517084 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.517462 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.536910 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lq95t\" (UniqueName: \"kubernetes.io/projected/4fe91a7c-36b7-4528-8061-f2c47bede56d-kube-api-access-lq95t\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-cwgpt\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:28 crc kubenswrapper[4787]: I1203 18:13:28.603982 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:13:29 crc kubenswrapper[4787]: I1203 18:13:29.202548 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt"] Dec 03 18:13:29 crc kubenswrapper[4787]: W1203 18:13:29.211371 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fe91a7c_36b7_4528_8061_f2c47bede56d.slice/crio-a0235fa18947723d94b963207d62ad5e890708b32b64b4d2c1a45974a89dae41 WatchSource:0}: Error finding container a0235fa18947723d94b963207d62ad5e890708b32b64b4d2c1a45974a89dae41: Status 404 returned error can't find the container with id a0235fa18947723d94b963207d62ad5e890708b32b64b4d2c1a45974a89dae41 Dec 03 18:13:30 crc kubenswrapper[4787]: I1203 18:13:30.233268 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" event={"ID":"4fe91a7c-36b7-4528-8061-f2c47bede56d","Type":"ContainerStarted","Data":"9198a3c81a912775861a36c1d8572b3a19f947c9db538e6ea6d0dbff117c989e"} Dec 03 18:13:30 crc kubenswrapper[4787]: I1203 18:13:30.233618 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" event={"ID":"4fe91a7c-36b7-4528-8061-f2c47bede56d","Type":"ContainerStarted","Data":"a0235fa18947723d94b963207d62ad5e890708b32b64b4d2c1a45974a89dae41"} Dec 03 18:13:30 crc kubenswrapper[4787]: I1203 18:13:30.256809 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" podStartSLOduration=1.855891353 podStartE2EDuration="2.256787668s" podCreationTimestamp="2025-12-03 18:13:28 +0000 UTC" firstStartedPulling="2025-12-03 18:13:29.214619951 +0000 UTC m=+3626.032090910" lastFinishedPulling="2025-12-03 18:13:29.615516266 +0000 UTC m=+3626.432987225" observedRunningTime="2025-12-03 18:13:30.24790272 +0000 UTC m=+3627.065373679" watchObservedRunningTime="2025-12-03 18:13:30.256787668 +0000 UTC m=+3627.074258617" Dec 03 18:13:48 crc kubenswrapper[4787]: I1203 18:13:48.990269 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:13:48 crc kubenswrapper[4787]: I1203 18:13:48.990886 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:13:48 crc kubenswrapper[4787]: I1203 18:13:48.990952 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 18:13:48 crc kubenswrapper[4787]: I1203 18:13:48.991886 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8027146e8048a2924f6ea539e853b86d156ce857b489636ed54a2f77398dc20e"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:13:48 crc kubenswrapper[4787]: I1203 18:13:48.991964 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://8027146e8048a2924f6ea539e853b86d156ce857b489636ed54a2f77398dc20e" gracePeriod=600 Dec 03 18:13:49 crc kubenswrapper[4787]: I1203 18:13:49.440529 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="8027146e8048a2924f6ea539e853b86d156ce857b489636ed54a2f77398dc20e" exitCode=0 Dec 03 18:13:49 crc kubenswrapper[4787]: I1203 18:13:49.440626 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"8027146e8048a2924f6ea539e853b86d156ce857b489636ed54a2f77398dc20e"} Dec 03 18:13:49 crc kubenswrapper[4787]: I1203 18:13:49.440929 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c"} Dec 03 18:13:49 crc kubenswrapper[4787]: I1203 18:13:49.440967 4787 scope.go:117] "RemoveContainer" containerID="dc65ec472fe23a290fd40291a35361e3bf67fc04e7015947a1192086da90adb8" Dec 03 18:14:49 crc kubenswrapper[4787]: I1203 18:14:49.067597 4787 generic.go:334] "Generic (PLEG): container finished" podID="4fe91a7c-36b7-4528-8061-f2c47bede56d" containerID="9198a3c81a912775861a36c1d8572b3a19f947c9db538e6ea6d0dbff117c989e" exitCode=0 Dec 03 18:14:49 crc kubenswrapper[4787]: I1203 18:14:49.067731 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" event={"ID":"4fe91a7c-36b7-4528-8061-f2c47bede56d","Type":"ContainerDied","Data":"9198a3c81a912775861a36c1d8572b3a19f947c9db538e6ea6d0dbff117c989e"} Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.632602 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.703708 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovn-combined-ca-bundle\") pod \"4fe91a7c-36b7-4528-8061-f2c47bede56d\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.703825 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ssh-key\") pod \"4fe91a7c-36b7-4528-8061-f2c47bede56d\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.703935 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lq95t\" (UniqueName: \"kubernetes.io/projected/4fe91a7c-36b7-4528-8061-f2c47bede56d-kube-api-access-lq95t\") pod \"4fe91a7c-36b7-4528-8061-f2c47bede56d\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.703971 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovncontroller-config-0\") pod \"4fe91a7c-36b7-4528-8061-f2c47bede56d\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.704003 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ceph\") pod \"4fe91a7c-36b7-4528-8061-f2c47bede56d\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.704231 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-inventory\") pod \"4fe91a7c-36b7-4528-8061-f2c47bede56d\" (UID: \"4fe91a7c-36b7-4528-8061-f2c47bede56d\") " Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.710031 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ceph" (OuterVolumeSpecName: "ceph") pod "4fe91a7c-36b7-4528-8061-f2c47bede56d" (UID: "4fe91a7c-36b7-4528-8061-f2c47bede56d"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.711205 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fe91a7c-36b7-4528-8061-f2c47bede56d-kube-api-access-lq95t" (OuterVolumeSpecName: "kube-api-access-lq95t") pod "4fe91a7c-36b7-4528-8061-f2c47bede56d" (UID: "4fe91a7c-36b7-4528-8061-f2c47bede56d"). InnerVolumeSpecName "kube-api-access-lq95t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.710696 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "4fe91a7c-36b7-4528-8061-f2c47bede56d" (UID: "4fe91a7c-36b7-4528-8061-f2c47bede56d"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.735942 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4fe91a7c-36b7-4528-8061-f2c47bede56d" (UID: "4fe91a7c-36b7-4528-8061-f2c47bede56d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.738362 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-inventory" (OuterVolumeSpecName: "inventory") pod "4fe91a7c-36b7-4528-8061-f2c47bede56d" (UID: "4fe91a7c-36b7-4528-8061-f2c47bede56d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.738743 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "4fe91a7c-36b7-4528-8061-f2c47bede56d" (UID: "4fe91a7c-36b7-4528-8061-f2c47bede56d"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.806191 4787 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.806222 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.806232 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lq95t\" (UniqueName: \"kubernetes.io/projected/4fe91a7c-36b7-4528-8061-f2c47bede56d-kube-api-access-lq95t\") on node \"crc\" DevicePath \"\"" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.806243 4787 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/4fe91a7c-36b7-4528-8061-f2c47bede56d-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.806253 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:14:50 crc kubenswrapper[4787]: I1203 18:14:50.806264 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fe91a7c-36b7-4528-8061-f2c47bede56d-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.089225 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" event={"ID":"4fe91a7c-36b7-4528-8061-f2c47bede56d","Type":"ContainerDied","Data":"a0235fa18947723d94b963207d62ad5e890708b32b64b4d2c1a45974a89dae41"} Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.089262 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a0235fa18947723d94b963207d62ad5e890708b32b64b4d2c1a45974a89dae41" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.089320 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-cwgpt" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.254547 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr"] Dec 03 18:14:51 crc kubenswrapper[4787]: E1203 18:14:51.254953 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fe91a7c-36b7-4528-8061-f2c47bede56d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.254968 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fe91a7c-36b7-4528-8061-f2c47bede56d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.255185 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fe91a7c-36b7-4528-8061-f2c47bede56d" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.255986 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.258671 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.258801 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.260328 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.260927 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.265442 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.265475 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.265537 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.274751 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr"] Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.418119 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.418165 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.418223 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.418816 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.418931 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-66g2k\" (UniqueName: \"kubernetes.io/projected/424ea689-3407-48b5-a7cc-63bced4b0565-kube-api-access-66g2k\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.419275 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.419337 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.522105 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.522174 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.522233 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.522407 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.522448 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-66g2k\" (UniqueName: \"kubernetes.io/projected/424ea689-3407-48b5-a7cc-63bced4b0565-kube-api-access-66g2k\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.522583 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.522617 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.529231 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.530070 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.530859 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.531726 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.533444 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.534164 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.552432 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-66g2k\" (UniqueName: \"kubernetes.io/projected/424ea689-3407-48b5-a7cc-63bced4b0565-kube-api-access-66g2k\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:51 crc kubenswrapper[4787]: I1203 18:14:51.574272 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:14:52 crc kubenswrapper[4787]: I1203 18:14:52.179254 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr"] Dec 03 18:14:53 crc kubenswrapper[4787]: I1203 18:14:53.110788 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" event={"ID":"424ea689-3407-48b5-a7cc-63bced4b0565","Type":"ContainerStarted","Data":"416228b5965ad9ada161ce20d90f61362a931f9cb99f3da417cd7453f8c2bf10"} Dec 03 18:14:53 crc kubenswrapper[4787]: I1203 18:14:53.111239 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" event={"ID":"424ea689-3407-48b5-a7cc-63bced4b0565","Type":"ContainerStarted","Data":"7f5c1bb6af5485e6e26603cc51fc371720c6bce6688a6c81bd5e61f04f2f4c18"} Dec 03 18:14:53 crc kubenswrapper[4787]: I1203 18:14:53.138779 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" podStartSLOduration=1.6414855 podStartE2EDuration="2.138757946s" podCreationTimestamp="2025-12-03 18:14:51 +0000 UTC" firstStartedPulling="2025-12-03 18:14:52.196236577 +0000 UTC m=+3709.013707536" lastFinishedPulling="2025-12-03 18:14:52.693509023 +0000 UTC m=+3709.510979982" observedRunningTime="2025-12-03 18:14:53.129049726 +0000 UTC m=+3709.946520715" watchObservedRunningTime="2025-12-03 18:14:53.138757946 +0000 UTC m=+3709.956228915" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.151692 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6"] Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.154648 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.162216 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.162400 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.172129 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6"] Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.317945 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08a749e7-ff3a-40e3-8a57-418cdda847b7-config-volume\") pod \"collect-profiles-29413095-cc9x6\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.318132 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08a749e7-ff3a-40e3-8a57-418cdda847b7-secret-volume\") pod \"collect-profiles-29413095-cc9x6\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.318217 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xlqn\" (UniqueName: \"kubernetes.io/projected/08a749e7-ff3a-40e3-8a57-418cdda847b7-kube-api-access-4xlqn\") pod \"collect-profiles-29413095-cc9x6\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.420458 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08a749e7-ff3a-40e3-8a57-418cdda847b7-secret-volume\") pod \"collect-profiles-29413095-cc9x6\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.420580 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xlqn\" (UniqueName: \"kubernetes.io/projected/08a749e7-ff3a-40e3-8a57-418cdda847b7-kube-api-access-4xlqn\") pod \"collect-profiles-29413095-cc9x6\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.420621 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08a749e7-ff3a-40e3-8a57-418cdda847b7-config-volume\") pod \"collect-profiles-29413095-cc9x6\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.421730 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08a749e7-ff3a-40e3-8a57-418cdda847b7-config-volume\") pod \"collect-profiles-29413095-cc9x6\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.436453 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08a749e7-ff3a-40e3-8a57-418cdda847b7-secret-volume\") pod \"collect-profiles-29413095-cc9x6\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.437224 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xlqn\" (UniqueName: \"kubernetes.io/projected/08a749e7-ff3a-40e3-8a57-418cdda847b7-kube-api-access-4xlqn\") pod \"collect-profiles-29413095-cc9x6\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.492807 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:00 crc kubenswrapper[4787]: I1203 18:15:00.949593 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6"] Dec 03 18:15:01 crc kubenswrapper[4787]: I1203 18:15:01.210646 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" event={"ID":"08a749e7-ff3a-40e3-8a57-418cdda847b7","Type":"ContainerStarted","Data":"60ad081957a6443a259c147724d6ecf88de75ccc4c3400a83ca52686ca64849a"} Dec 03 18:15:01 crc kubenswrapper[4787]: I1203 18:15:01.210959 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" event={"ID":"08a749e7-ff3a-40e3-8a57-418cdda847b7","Type":"ContainerStarted","Data":"cc1c7a58d79b41a5de88f698d6016623708215fc824b8d9b83fbe38f7030095d"} Dec 03 18:15:01 crc kubenswrapper[4787]: I1203 18:15:01.243317 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" podStartSLOduration=1.243298275 podStartE2EDuration="1.243298275s" podCreationTimestamp="2025-12-03 18:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:15:01.225509039 +0000 UTC m=+3718.042979988" watchObservedRunningTime="2025-12-03 18:15:01.243298275 +0000 UTC m=+3718.060769234" Dec 03 18:15:02 crc kubenswrapper[4787]: I1203 18:15:02.219650 4787 generic.go:334] "Generic (PLEG): container finished" podID="08a749e7-ff3a-40e3-8a57-418cdda847b7" containerID="60ad081957a6443a259c147724d6ecf88de75ccc4c3400a83ca52686ca64849a" exitCode=0 Dec 03 18:15:02 crc kubenswrapper[4787]: I1203 18:15:02.219694 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" event={"ID":"08a749e7-ff3a-40e3-8a57-418cdda847b7","Type":"ContainerDied","Data":"60ad081957a6443a259c147724d6ecf88de75ccc4c3400a83ca52686ca64849a"} Dec 03 18:15:03 crc kubenswrapper[4787]: I1203 18:15:03.669060 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:03 crc kubenswrapper[4787]: I1203 18:15:03.793843 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08a749e7-ff3a-40e3-8a57-418cdda847b7-secret-volume\") pod \"08a749e7-ff3a-40e3-8a57-418cdda847b7\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " Dec 03 18:15:03 crc kubenswrapper[4787]: I1203 18:15:03.793981 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08a749e7-ff3a-40e3-8a57-418cdda847b7-config-volume\") pod \"08a749e7-ff3a-40e3-8a57-418cdda847b7\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " Dec 03 18:15:03 crc kubenswrapper[4787]: I1203 18:15:03.794043 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4xlqn\" (UniqueName: \"kubernetes.io/projected/08a749e7-ff3a-40e3-8a57-418cdda847b7-kube-api-access-4xlqn\") pod \"08a749e7-ff3a-40e3-8a57-418cdda847b7\" (UID: \"08a749e7-ff3a-40e3-8a57-418cdda847b7\") " Dec 03 18:15:03 crc kubenswrapper[4787]: I1203 18:15:03.801948 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/08a749e7-ff3a-40e3-8a57-418cdda847b7-config-volume" (OuterVolumeSpecName: "config-volume") pod "08a749e7-ff3a-40e3-8a57-418cdda847b7" (UID: "08a749e7-ff3a-40e3-8a57-418cdda847b7"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:15:03 crc kubenswrapper[4787]: I1203 18:15:03.829149 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/08a749e7-ff3a-40e3-8a57-418cdda847b7-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "08a749e7-ff3a-40e3-8a57-418cdda847b7" (UID: "08a749e7-ff3a-40e3-8a57-418cdda847b7"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:15:03 crc kubenswrapper[4787]: I1203 18:15:03.832222 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/08a749e7-ff3a-40e3-8a57-418cdda847b7-kube-api-access-4xlqn" (OuterVolumeSpecName: "kube-api-access-4xlqn") pod "08a749e7-ff3a-40e3-8a57-418cdda847b7" (UID: "08a749e7-ff3a-40e3-8a57-418cdda847b7"). InnerVolumeSpecName "kube-api-access-4xlqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:15:03 crc kubenswrapper[4787]: I1203 18:15:03.896898 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/08a749e7-ff3a-40e3-8a57-418cdda847b7-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:03 crc kubenswrapper[4787]: I1203 18:15:03.897163 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/08a749e7-ff3a-40e3-8a57-418cdda847b7-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:03 crc kubenswrapper[4787]: I1203 18:15:03.897173 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4xlqn\" (UniqueName: \"kubernetes.io/projected/08a749e7-ff3a-40e3-8a57-418cdda847b7-kube-api-access-4xlqn\") on node \"crc\" DevicePath \"\"" Dec 03 18:15:04 crc kubenswrapper[4787]: I1203 18:15:04.243506 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" event={"ID":"08a749e7-ff3a-40e3-8a57-418cdda847b7","Type":"ContainerDied","Data":"cc1c7a58d79b41a5de88f698d6016623708215fc824b8d9b83fbe38f7030095d"} Dec 03 18:15:04 crc kubenswrapper[4787]: I1203 18:15:04.243543 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cc1c7a58d79b41a5de88f698d6016623708215fc824b8d9b83fbe38f7030095d" Dec 03 18:15:04 crc kubenswrapper[4787]: I1203 18:15:04.243544 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6" Dec 03 18:15:04 crc kubenswrapper[4787]: I1203 18:15:04.316884 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d"] Dec 03 18:15:04 crc kubenswrapper[4787]: I1203 18:15:04.325201 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413050-jgj7d"] Dec 03 18:15:05 crc kubenswrapper[4787]: I1203 18:15:05.787382 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a5b1845b-dcdb-4fb2-983a-4018aa964656" path="/var/lib/kubelet/pods/a5b1845b-dcdb-4fb2-983a-4018aa964656/volumes" Dec 03 18:15:11 crc kubenswrapper[4787]: E1203 18:15:11.183928 4787 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="5.418s" Dec 03 18:15:26 crc kubenswrapper[4787]: I1203 18:15:26.858876 4787 scope.go:117] "RemoveContainer" containerID="017e800dc669ff109cf880c14e980f50c5d5c95a101d8fb12f9a6a7631d30b85" Dec 03 18:16:18 crc kubenswrapper[4787]: I1203 18:16:18.989911 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:16:18 crc kubenswrapper[4787]: I1203 18:16:18.990475 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:16:22 crc kubenswrapper[4787]: I1203 18:16:22.119982 4787 generic.go:334] "Generic (PLEG): container finished" podID="424ea689-3407-48b5-a7cc-63bced4b0565" containerID="416228b5965ad9ada161ce20d90f61362a931f9cb99f3da417cd7453f8c2bf10" exitCode=0 Dec 03 18:16:22 crc kubenswrapper[4787]: I1203 18:16:22.120082 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" event={"ID":"424ea689-3407-48b5-a7cc-63bced4b0565","Type":"ContainerDied","Data":"416228b5965ad9ada161ce20d90f61362a931f9cb99f3da417cd7453f8c2bf10"} Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.646077 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.699441 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-inventory\") pod \"424ea689-3407-48b5-a7cc-63bced4b0565\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.699505 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ceph\") pod \"424ea689-3407-48b5-a7cc-63bced4b0565\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.699593 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-nova-metadata-neutron-config-0\") pod \"424ea689-3407-48b5-a7cc-63bced4b0565\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.699677 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-66g2k\" (UniqueName: \"kubernetes.io/projected/424ea689-3407-48b5-a7cc-63bced4b0565-kube-api-access-66g2k\") pod \"424ea689-3407-48b5-a7cc-63bced4b0565\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.699734 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-ovn-metadata-agent-neutron-config-0\") pod \"424ea689-3407-48b5-a7cc-63bced4b0565\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.699757 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ssh-key\") pod \"424ea689-3407-48b5-a7cc-63bced4b0565\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.699865 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-metadata-combined-ca-bundle\") pod \"424ea689-3407-48b5-a7cc-63bced4b0565\" (UID: \"424ea689-3407-48b5-a7cc-63bced4b0565\") " Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.706157 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "424ea689-3407-48b5-a7cc-63bced4b0565" (UID: "424ea689-3407-48b5-a7cc-63bced4b0565"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.706232 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ceph" (OuterVolumeSpecName: "ceph") pod "424ea689-3407-48b5-a7cc-63bced4b0565" (UID: "424ea689-3407-48b5-a7cc-63bced4b0565"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.706265 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/424ea689-3407-48b5-a7cc-63bced4b0565-kube-api-access-66g2k" (OuterVolumeSpecName: "kube-api-access-66g2k") pod "424ea689-3407-48b5-a7cc-63bced4b0565" (UID: "424ea689-3407-48b5-a7cc-63bced4b0565"). InnerVolumeSpecName "kube-api-access-66g2k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.736924 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "424ea689-3407-48b5-a7cc-63bced4b0565" (UID: "424ea689-3407-48b5-a7cc-63bced4b0565"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.738029 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "424ea689-3407-48b5-a7cc-63bced4b0565" (UID: "424ea689-3407-48b5-a7cc-63bced4b0565"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.744151 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "424ea689-3407-48b5-a7cc-63bced4b0565" (UID: "424ea689-3407-48b5-a7cc-63bced4b0565"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.753889 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-inventory" (OuterVolumeSpecName: "inventory") pod "424ea689-3407-48b5-a7cc-63bced4b0565" (UID: "424ea689-3407-48b5-a7cc-63bced4b0565"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.804181 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.804229 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.804248 4787 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.804269 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-66g2k\" (UniqueName: \"kubernetes.io/projected/424ea689-3407-48b5-a7cc-63bced4b0565-kube-api-access-66g2k\") on node \"crc\" DevicePath \"\"" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.804290 4787 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.804309 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:16:23 crc kubenswrapper[4787]: I1203 18:16:23.804328 4787 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/424ea689-3407-48b5-a7cc-63bced4b0565-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.144539 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" event={"ID":"424ea689-3407-48b5-a7cc-63bced4b0565","Type":"ContainerDied","Data":"7f5c1bb6af5485e6e26603cc51fc371720c6bce6688a6c81bd5e61f04f2f4c18"} Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.144887 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f5c1bb6af5485e6e26603cc51fc371720c6bce6688a6c81bd5e61f04f2f4c18" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.144588 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.257313 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs"] Dec 03 18:16:24 crc kubenswrapper[4787]: E1203 18:16:24.257900 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="424ea689-3407-48b5-a7cc-63bced4b0565" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.257924 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="424ea689-3407-48b5-a7cc-63bced4b0565" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 18:16:24 crc kubenswrapper[4787]: E1203 18:16:24.257967 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="08a749e7-ff3a-40e3-8a57-418cdda847b7" containerName="collect-profiles" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.257974 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="08a749e7-ff3a-40e3-8a57-418cdda847b7" containerName="collect-profiles" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.258226 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="424ea689-3407-48b5-a7cc-63bced4b0565" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.258246 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="08a749e7-ff3a-40e3-8a57-418cdda847b7" containerName="collect-profiles" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.259048 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.262498 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.263460 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.263692 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.263862 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.266049 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.268121 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.272621 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs"] Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.312946 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.313102 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.313193 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.313240 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.313274 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.313298 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d7v5q\" (UniqueName: \"kubernetes.io/projected/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-kube-api-access-d7v5q\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.415287 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.415404 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.415461 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.415502 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.415534 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.415555 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d7v5q\" (UniqueName: \"kubernetes.io/projected/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-kube-api-access-d7v5q\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.419550 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.420116 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.420119 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.422226 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.422337 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.438432 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d7v5q\" (UniqueName: \"kubernetes.io/projected/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-kube-api-access-d7v5q\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:24 crc kubenswrapper[4787]: I1203 18:16:24.584266 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:16:25 crc kubenswrapper[4787]: I1203 18:16:25.166766 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs"] Dec 03 18:16:26 crc kubenswrapper[4787]: I1203 18:16:26.169371 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" event={"ID":"7acc721f-ca1b-4a7c-9e11-4c3b58096a51","Type":"ContainerStarted","Data":"8db50fce78ec3d4846684ad2607dbb8545b54bc27aedf43a2e87ba4686eaa4b3"} Dec 03 18:16:26 crc kubenswrapper[4787]: I1203 18:16:26.169789 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" event={"ID":"7acc721f-ca1b-4a7c-9e11-4c3b58096a51","Type":"ContainerStarted","Data":"0cb11df18cb2d52c3cfb81018cef842aac345b525b2d2849177cd6f8a8f7908f"} Dec 03 18:16:26 crc kubenswrapper[4787]: I1203 18:16:26.196995 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" podStartSLOduration=1.748940916 podStartE2EDuration="2.196980943s" podCreationTimestamp="2025-12-03 18:16:24 +0000 UTC" firstStartedPulling="2025-12-03 18:16:25.168368929 +0000 UTC m=+3801.985839888" lastFinishedPulling="2025-12-03 18:16:25.616408956 +0000 UTC m=+3802.433879915" observedRunningTime="2025-12-03 18:16:26.194097515 +0000 UTC m=+3803.011568544" watchObservedRunningTime="2025-12-03 18:16:26.196980943 +0000 UTC m=+3803.014451902" Dec 03 18:16:48 crc kubenswrapper[4787]: I1203 18:16:48.990006 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:16:48 crc kubenswrapper[4787]: I1203 18:16:48.990593 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:17:18 crc kubenswrapper[4787]: I1203 18:17:18.989812 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:17:18 crc kubenswrapper[4787]: I1203 18:17:18.990354 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:17:18 crc kubenswrapper[4787]: I1203 18:17:18.990399 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 18:17:18 crc kubenswrapper[4787]: I1203 18:17:18.991245 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:17:18 crc kubenswrapper[4787]: I1203 18:17:18.991301 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" gracePeriod=600 Dec 03 18:17:19 crc kubenswrapper[4787]: E1203 18:17:19.140812 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:17:19 crc kubenswrapper[4787]: I1203 18:17:19.803651 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" exitCode=0 Dec 03 18:17:19 crc kubenswrapper[4787]: I1203 18:17:19.803750 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c"} Dec 03 18:17:19 crc kubenswrapper[4787]: I1203 18:17:19.804186 4787 scope.go:117] "RemoveContainer" containerID="8027146e8048a2924f6ea539e853b86d156ce857b489636ed54a2f77398dc20e" Dec 03 18:17:19 crc kubenswrapper[4787]: I1203 18:17:19.805560 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:17:19 crc kubenswrapper[4787]: E1203 18:17:19.809201 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:17:30 crc kubenswrapper[4787]: I1203 18:17:30.767250 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:17:30 crc kubenswrapper[4787]: E1203 18:17:30.768315 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:17:43 crc kubenswrapper[4787]: I1203 18:17:43.786271 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:17:43 crc kubenswrapper[4787]: E1203 18:17:43.787767 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:17:54 crc kubenswrapper[4787]: I1203 18:17:54.766988 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:17:54 crc kubenswrapper[4787]: E1203 18:17:54.767939 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:18:06 crc kubenswrapper[4787]: I1203 18:18:06.767264 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:18:06 crc kubenswrapper[4787]: E1203 18:18:06.768038 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:18:19 crc kubenswrapper[4787]: I1203 18:18:19.766917 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:18:19 crc kubenswrapper[4787]: E1203 18:18:19.767823 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:18:30 crc kubenswrapper[4787]: I1203 18:18:30.766974 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:18:30 crc kubenswrapper[4787]: E1203 18:18:30.768206 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:18:42 crc kubenswrapper[4787]: I1203 18:18:42.766429 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:18:42 crc kubenswrapper[4787]: E1203 18:18:42.767607 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:18:54 crc kubenswrapper[4787]: I1203 18:18:54.766357 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:18:54 crc kubenswrapper[4787]: E1203 18:18:54.767080 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.028943 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9tmqh"] Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.031582 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.042950 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9tmqh"] Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.049904 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-utilities\") pod \"community-operators-9tmqh\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.050002 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r22qm\" (UniqueName: \"kubernetes.io/projected/212489b3-1bfd-44a0-8c31-358d3d20c5a1-kube-api-access-r22qm\") pod \"community-operators-9tmqh\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.050101 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-catalog-content\") pod \"community-operators-9tmqh\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.152222 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-utilities\") pod \"community-operators-9tmqh\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.152287 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r22qm\" (UniqueName: \"kubernetes.io/projected/212489b3-1bfd-44a0-8c31-358d3d20c5a1-kube-api-access-r22qm\") pod \"community-operators-9tmqh\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.152333 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-catalog-content\") pod \"community-operators-9tmqh\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.152701 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-utilities\") pod \"community-operators-9tmqh\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.152726 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-catalog-content\") pod \"community-operators-9tmqh\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.174697 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r22qm\" (UniqueName: \"kubernetes.io/projected/212489b3-1bfd-44a0-8c31-358d3d20c5a1-kube-api-access-r22qm\") pod \"community-operators-9tmqh\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.361116 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:18:55 crc kubenswrapper[4787]: I1203 18:18:55.933912 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9tmqh"] Dec 03 18:18:56 crc kubenswrapper[4787]: I1203 18:18:56.918827 4787 generic.go:334] "Generic (PLEG): container finished" podID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerID="80b0575c4b76bfb608c6f00d270b7e1fb4b30c5ccddedcc5b0b6eb693aa339b6" exitCode=0 Dec 03 18:18:56 crc kubenswrapper[4787]: I1203 18:18:56.918926 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tmqh" event={"ID":"212489b3-1bfd-44a0-8c31-358d3d20c5a1","Type":"ContainerDied","Data":"80b0575c4b76bfb608c6f00d270b7e1fb4b30c5ccddedcc5b0b6eb693aa339b6"} Dec 03 18:18:56 crc kubenswrapper[4787]: I1203 18:18:56.919066 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tmqh" event={"ID":"212489b3-1bfd-44a0-8c31-358d3d20c5a1","Type":"ContainerStarted","Data":"9048fd5d3f99e7b52c6e2f86d10bb1a1a20c0570bb0f48540e5c3ba3498ade85"} Dec 03 18:18:56 crc kubenswrapper[4787]: I1203 18:18:56.920910 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:18:58 crc kubenswrapper[4787]: I1203 18:18:58.937947 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tmqh" event={"ID":"212489b3-1bfd-44a0-8c31-358d3d20c5a1","Type":"ContainerStarted","Data":"5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698"} Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.017883 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mlxfc"] Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.021931 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.060232 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mlxfc"] Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.178526 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-utilities\") pod \"certified-operators-mlxfc\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.178792 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f2np6\" (UniqueName: \"kubernetes.io/projected/3a1d2603-4d44-48ff-b89a-f0fec82572fd-kube-api-access-f2np6\") pod \"certified-operators-mlxfc\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.178869 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-catalog-content\") pod \"certified-operators-mlxfc\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.281227 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f2np6\" (UniqueName: \"kubernetes.io/projected/3a1d2603-4d44-48ff-b89a-f0fec82572fd-kube-api-access-f2np6\") pod \"certified-operators-mlxfc\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.281322 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-catalog-content\") pod \"certified-operators-mlxfc\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.281397 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-utilities\") pod \"certified-operators-mlxfc\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.282056 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-utilities\") pod \"certified-operators-mlxfc\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.282700 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-catalog-content\") pod \"certified-operators-mlxfc\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.307218 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f2np6\" (UniqueName: \"kubernetes.io/projected/3a1d2603-4d44-48ff-b89a-f0fec82572fd-kube-api-access-f2np6\") pod \"certified-operators-mlxfc\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.355366 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.861052 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mlxfc"] Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.973052 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlxfc" event={"ID":"3a1d2603-4d44-48ff-b89a-f0fec82572fd","Type":"ContainerStarted","Data":"879090e1cde22075af2422d6972b4188f506025952c989d50bda9c56cf40348c"} Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.975311 4787 generic.go:334] "Generic (PLEG): container finished" podID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerID="5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698" exitCode=0 Dec 03 18:19:01 crc kubenswrapper[4787]: I1203 18:19:01.975339 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tmqh" event={"ID":"212489b3-1bfd-44a0-8c31-358d3d20c5a1","Type":"ContainerDied","Data":"5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698"} Dec 03 18:19:02 crc kubenswrapper[4787]: I1203 18:19:02.988916 4787 generic.go:334] "Generic (PLEG): container finished" podID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" containerID="a3741a01699dabd462f4056840ab719825f9afcf3a9bfd7ae9b1c19910a98f5b" exitCode=0 Dec 03 18:19:02 crc kubenswrapper[4787]: I1203 18:19:02.989235 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlxfc" event={"ID":"3a1d2603-4d44-48ff-b89a-f0fec82572fd","Type":"ContainerDied","Data":"a3741a01699dabd462f4056840ab719825f9afcf3a9bfd7ae9b1c19910a98f5b"} Dec 03 18:19:02 crc kubenswrapper[4787]: I1203 18:19:02.997891 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tmqh" event={"ID":"212489b3-1bfd-44a0-8c31-358d3d20c5a1","Type":"ContainerStarted","Data":"6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39"} Dec 03 18:19:03 crc kubenswrapper[4787]: I1203 18:19:03.037155 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9tmqh" podStartSLOduration=2.557844007 podStartE2EDuration="8.03713487s" podCreationTimestamp="2025-12-03 18:18:55 +0000 UTC" firstStartedPulling="2025-12-03 18:18:56.920679595 +0000 UTC m=+3953.738150554" lastFinishedPulling="2025-12-03 18:19:02.399970458 +0000 UTC m=+3959.217441417" observedRunningTime="2025-12-03 18:19:03.031759656 +0000 UTC m=+3959.849230625" watchObservedRunningTime="2025-12-03 18:19:03.03713487 +0000 UTC m=+3959.854605829" Dec 03 18:19:05 crc kubenswrapper[4787]: I1203 18:19:05.021298 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlxfc" event={"ID":"3a1d2603-4d44-48ff-b89a-f0fec82572fd","Type":"ContainerStarted","Data":"adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f"} Dec 03 18:19:05 crc kubenswrapper[4787]: I1203 18:19:05.362356 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:19:05 crc kubenswrapper[4787]: I1203 18:19:05.362678 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:19:06 crc kubenswrapper[4787]: I1203 18:19:06.031979 4787 generic.go:334] "Generic (PLEG): container finished" podID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" containerID="adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f" exitCode=0 Dec 03 18:19:06 crc kubenswrapper[4787]: I1203 18:19:06.032051 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlxfc" event={"ID":"3a1d2603-4d44-48ff-b89a-f0fec82572fd","Type":"ContainerDied","Data":"adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f"} Dec 03 18:19:06 crc kubenswrapper[4787]: I1203 18:19:06.413464 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-9tmqh" podUID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerName="registry-server" probeResult="failure" output=< Dec 03 18:19:06 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 18:19:06 crc kubenswrapper[4787]: > Dec 03 18:19:08 crc kubenswrapper[4787]: I1203 18:19:08.055602 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlxfc" event={"ID":"3a1d2603-4d44-48ff-b89a-f0fec82572fd","Type":"ContainerStarted","Data":"6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c"} Dec 03 18:19:08 crc kubenswrapper[4787]: I1203 18:19:08.105454 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mlxfc" podStartSLOduration=4.35618153 podStartE2EDuration="8.105430296s" podCreationTimestamp="2025-12-03 18:19:00 +0000 UTC" firstStartedPulling="2025-12-03 18:19:02.991265432 +0000 UTC m=+3959.808736391" lastFinishedPulling="2025-12-03 18:19:06.740514188 +0000 UTC m=+3963.557985157" observedRunningTime="2025-12-03 18:19:08.071257501 +0000 UTC m=+3964.888728490" watchObservedRunningTime="2025-12-03 18:19:08.105430296 +0000 UTC m=+3964.922901255" Dec 03 18:19:08 crc kubenswrapper[4787]: I1203 18:19:08.766629 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:19:08 crc kubenswrapper[4787]: E1203 18:19:08.767125 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:19:11 crc kubenswrapper[4787]: I1203 18:19:11.356142 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:11 crc kubenswrapper[4787]: I1203 18:19:11.356502 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:11 crc kubenswrapper[4787]: I1203 18:19:11.424431 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:12 crc kubenswrapper[4787]: I1203 18:19:12.155975 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:13 crc kubenswrapper[4787]: I1203 18:19:13.012317 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mlxfc"] Dec 03 18:19:14 crc kubenswrapper[4787]: I1203 18:19:14.109754 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-mlxfc" podUID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" containerName="registry-server" containerID="cri-o://6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c" gracePeriod=2 Dec 03 18:19:14 crc kubenswrapper[4787]: I1203 18:19:14.729563 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:14 crc kubenswrapper[4787]: I1203 18:19:14.774620 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-utilities\") pod \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " Dec 03 18:19:14 crc kubenswrapper[4787]: I1203 18:19:14.774724 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f2np6\" (UniqueName: \"kubernetes.io/projected/3a1d2603-4d44-48ff-b89a-f0fec82572fd-kube-api-access-f2np6\") pod \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " Dec 03 18:19:14 crc kubenswrapper[4787]: I1203 18:19:14.774801 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-catalog-content\") pod \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\" (UID: \"3a1d2603-4d44-48ff-b89a-f0fec82572fd\") " Dec 03 18:19:14 crc kubenswrapper[4787]: I1203 18:19:14.775418 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-utilities" (OuterVolumeSpecName: "utilities") pod "3a1d2603-4d44-48ff-b89a-f0fec82572fd" (UID: "3a1d2603-4d44-48ff-b89a-f0fec82572fd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:19:14 crc kubenswrapper[4787]: I1203 18:19:14.790841 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a1d2603-4d44-48ff-b89a-f0fec82572fd-kube-api-access-f2np6" (OuterVolumeSpecName: "kube-api-access-f2np6") pod "3a1d2603-4d44-48ff-b89a-f0fec82572fd" (UID: "3a1d2603-4d44-48ff-b89a-f0fec82572fd"). InnerVolumeSpecName "kube-api-access-f2np6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:19:14 crc kubenswrapper[4787]: I1203 18:19:14.826311 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a1d2603-4d44-48ff-b89a-f0fec82572fd" (UID: "3a1d2603-4d44-48ff-b89a-f0fec82572fd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:19:14 crc kubenswrapper[4787]: I1203 18:19:14.877436 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:14 crc kubenswrapper[4787]: I1203 18:19:14.877473 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f2np6\" (UniqueName: \"kubernetes.io/projected/3a1d2603-4d44-48ff-b89a-f0fec82572fd-kube-api-access-f2np6\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:14 crc kubenswrapper[4787]: I1203 18:19:14.877486 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a1d2603-4d44-48ff-b89a-f0fec82572fd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.120420 4787 generic.go:334] "Generic (PLEG): container finished" podID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" containerID="6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c" exitCode=0 Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.120459 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlxfc" event={"ID":"3a1d2603-4d44-48ff-b89a-f0fec82572fd","Type":"ContainerDied","Data":"6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c"} Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.120486 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mlxfc" event={"ID":"3a1d2603-4d44-48ff-b89a-f0fec82572fd","Type":"ContainerDied","Data":"879090e1cde22075af2422d6972b4188f506025952c989d50bda9c56cf40348c"} Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.120496 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mlxfc" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.120502 4787 scope.go:117] "RemoveContainer" containerID="6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.152702 4787 scope.go:117] "RemoveContainer" containerID="adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.172536 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-mlxfc"] Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.184235 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-mlxfc"] Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.188940 4787 scope.go:117] "RemoveContainer" containerID="a3741a01699dabd462f4056840ab719825f9afcf3a9bfd7ae9b1c19910a98f5b" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.231576 4787 scope.go:117] "RemoveContainer" containerID="6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c" Dec 03 18:19:15 crc kubenswrapper[4787]: E1203 18:19:15.232544 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c\": container with ID starting with 6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c not found: ID does not exist" containerID="6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.232597 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c"} err="failed to get container status \"6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c\": rpc error: code = NotFound desc = could not find container \"6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c\": container with ID starting with 6df80e99fda076c6ae2995019e31de6e65b3e9f9ae509fbedd72b6696615799c not found: ID does not exist" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.232628 4787 scope.go:117] "RemoveContainer" containerID="adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f" Dec 03 18:19:15 crc kubenswrapper[4787]: E1203 18:19:15.233100 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f\": container with ID starting with adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f not found: ID does not exist" containerID="adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.233141 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f"} err="failed to get container status \"adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f\": rpc error: code = NotFound desc = could not find container \"adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f\": container with ID starting with adad78ac398a4cffd94908f6de558601ba8d7f349ebecde299d3dfcf6560558f not found: ID does not exist" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.233165 4787 scope.go:117] "RemoveContainer" containerID="a3741a01699dabd462f4056840ab719825f9afcf3a9bfd7ae9b1c19910a98f5b" Dec 03 18:19:15 crc kubenswrapper[4787]: E1203 18:19:15.233500 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3741a01699dabd462f4056840ab719825f9afcf3a9bfd7ae9b1c19910a98f5b\": container with ID starting with a3741a01699dabd462f4056840ab719825f9afcf3a9bfd7ae9b1c19910a98f5b not found: ID does not exist" containerID="a3741a01699dabd462f4056840ab719825f9afcf3a9bfd7ae9b1c19910a98f5b" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.233534 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3741a01699dabd462f4056840ab719825f9afcf3a9bfd7ae9b1c19910a98f5b"} err="failed to get container status \"a3741a01699dabd462f4056840ab719825f9afcf3a9bfd7ae9b1c19910a98f5b\": rpc error: code = NotFound desc = could not find container \"a3741a01699dabd462f4056840ab719825f9afcf3a9bfd7ae9b1c19910a98f5b\": container with ID starting with a3741a01699dabd462f4056840ab719825f9afcf3a9bfd7ae9b1c19910a98f5b not found: ID does not exist" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.432745 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.510705 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:19:15 crc kubenswrapper[4787]: I1203 18:19:15.803451 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" path="/var/lib/kubelet/pods/3a1d2603-4d44-48ff-b89a-f0fec82572fd/volumes" Dec 03 18:19:19 crc kubenswrapper[4787]: I1203 18:19:19.765868 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:19:19 crc kubenswrapper[4787]: E1203 18:19:19.766615 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:19:20 crc kubenswrapper[4787]: I1203 18:19:20.217378 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9tmqh"] Dec 03 18:19:20 crc kubenswrapper[4787]: I1203 18:19:20.217635 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9tmqh" podUID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerName="registry-server" containerID="cri-o://6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39" gracePeriod=2 Dec 03 18:19:20 crc kubenswrapper[4787]: E1203 18:19:20.331801 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod212489b3_1bfd_44a0_8c31_358d3d20c5a1.slice/crio-conmon-6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39.scope\": RecentStats: unable to find data in memory cache]" Dec 03 18:19:20 crc kubenswrapper[4787]: I1203 18:19:20.836006 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.021665 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-catalog-content\") pod \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.022199 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-utilities\") pod \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.022292 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r22qm\" (UniqueName: \"kubernetes.io/projected/212489b3-1bfd-44a0-8c31-358d3d20c5a1-kube-api-access-r22qm\") pod \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\" (UID: \"212489b3-1bfd-44a0-8c31-358d3d20c5a1\") " Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.022729 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-utilities" (OuterVolumeSpecName: "utilities") pod "212489b3-1bfd-44a0-8c31-358d3d20c5a1" (UID: "212489b3-1bfd-44a0-8c31-358d3d20c5a1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.023219 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.030416 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/212489b3-1bfd-44a0-8c31-358d3d20c5a1-kube-api-access-r22qm" (OuterVolumeSpecName: "kube-api-access-r22qm") pod "212489b3-1bfd-44a0-8c31-358d3d20c5a1" (UID: "212489b3-1bfd-44a0-8c31-358d3d20c5a1"). InnerVolumeSpecName "kube-api-access-r22qm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.074198 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "212489b3-1bfd-44a0-8c31-358d3d20c5a1" (UID: "212489b3-1bfd-44a0-8c31-358d3d20c5a1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.124945 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/212489b3-1bfd-44a0-8c31-358d3d20c5a1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.124979 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r22qm\" (UniqueName: \"kubernetes.io/projected/212489b3-1bfd-44a0-8c31-358d3d20c5a1-kube-api-access-r22qm\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.201205 4787 generic.go:334] "Generic (PLEG): container finished" podID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerID="6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39" exitCode=0 Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.201247 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tmqh" event={"ID":"212489b3-1bfd-44a0-8c31-358d3d20c5a1","Type":"ContainerDied","Data":"6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39"} Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.201255 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9tmqh" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.201280 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9tmqh" event={"ID":"212489b3-1bfd-44a0-8c31-358d3d20c5a1","Type":"ContainerDied","Data":"9048fd5d3f99e7b52c6e2f86d10bb1a1a20c0570bb0f48540e5c3ba3498ade85"} Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.201299 4787 scope.go:117] "RemoveContainer" containerID="6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.241489 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9tmqh"] Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.251714 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9tmqh"] Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.783664 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" path="/var/lib/kubelet/pods/212489b3-1bfd-44a0-8c31-358d3d20c5a1/volumes" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.904572 4787 scope.go:117] "RemoveContainer" containerID="5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.947444 4787 scope.go:117] "RemoveContainer" containerID="80b0575c4b76bfb608c6f00d270b7e1fb4b30c5ccddedcc5b0b6eb693aa339b6" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.972257 4787 scope.go:117] "RemoveContainer" containerID="6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39" Dec 03 18:19:21 crc kubenswrapper[4787]: E1203 18:19:21.972805 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39\": container with ID starting with 6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39 not found: ID does not exist" containerID="6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.972855 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39"} err="failed to get container status \"6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39\": rpc error: code = NotFound desc = could not find container \"6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39\": container with ID starting with 6b9c8fab6ea6d4d56a916ae3c95be4d5e090f6cda38685256d0b38a9131dac39 not found: ID does not exist" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.972887 4787 scope.go:117] "RemoveContainer" containerID="5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698" Dec 03 18:19:21 crc kubenswrapper[4787]: E1203 18:19:21.973348 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698\": container with ID starting with 5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698 not found: ID does not exist" containerID="5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.973454 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698"} err="failed to get container status \"5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698\": rpc error: code = NotFound desc = could not find container \"5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698\": container with ID starting with 5905247f39ef739174e9b5843b284e6b775e6402765d26629f3bf617c1938698 not found: ID does not exist" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.973542 4787 scope.go:117] "RemoveContainer" containerID="80b0575c4b76bfb608c6f00d270b7e1fb4b30c5ccddedcc5b0b6eb693aa339b6" Dec 03 18:19:21 crc kubenswrapper[4787]: E1203 18:19:21.974036 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"80b0575c4b76bfb608c6f00d270b7e1fb4b30c5ccddedcc5b0b6eb693aa339b6\": container with ID starting with 80b0575c4b76bfb608c6f00d270b7e1fb4b30c5ccddedcc5b0b6eb693aa339b6 not found: ID does not exist" containerID="80b0575c4b76bfb608c6f00d270b7e1fb4b30c5ccddedcc5b0b6eb693aa339b6" Dec 03 18:19:21 crc kubenswrapper[4787]: I1203 18:19:21.974087 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"80b0575c4b76bfb608c6f00d270b7e1fb4b30c5ccddedcc5b0b6eb693aa339b6"} err="failed to get container status \"80b0575c4b76bfb608c6f00d270b7e1fb4b30c5ccddedcc5b0b6eb693aa339b6\": rpc error: code = NotFound desc = could not find container \"80b0575c4b76bfb608c6f00d270b7e1fb4b30c5ccddedcc5b0b6eb693aa339b6\": container with ID starting with 80b0575c4b76bfb608c6f00d270b7e1fb4b30c5ccddedcc5b0b6eb693aa339b6 not found: ID does not exist" Dec 03 18:19:33 crc kubenswrapper[4787]: I1203 18:19:33.772750 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:19:33 crc kubenswrapper[4787]: E1203 18:19:33.773811 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:19:34 crc kubenswrapper[4787]: I1203 18:19:34.358776 4787 generic.go:334] "Generic (PLEG): container finished" podID="7acc721f-ca1b-4a7c-9e11-4c3b58096a51" containerID="8db50fce78ec3d4846684ad2607dbb8545b54bc27aedf43a2e87ba4686eaa4b3" exitCode=0 Dec 03 18:19:34 crc kubenswrapper[4787]: I1203 18:19:34.358839 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" event={"ID":"7acc721f-ca1b-4a7c-9e11-4c3b58096a51","Type":"ContainerDied","Data":"8db50fce78ec3d4846684ad2607dbb8545b54bc27aedf43a2e87ba4686eaa4b3"} Dec 03 18:19:35 crc kubenswrapper[4787]: I1203 18:19:35.851609 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.001317 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d7v5q\" (UniqueName: \"kubernetes.io/projected/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-kube-api-access-d7v5q\") pod \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.001554 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ceph\") pod \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.001629 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ssh-key\") pod \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.001681 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-secret-0\") pod \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.001710 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-combined-ca-bundle\") pod \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.001759 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-inventory\") pod \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\" (UID: \"7acc721f-ca1b-4a7c-9e11-4c3b58096a51\") " Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.008903 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-kube-api-access-d7v5q" (OuterVolumeSpecName: "kube-api-access-d7v5q") pod "7acc721f-ca1b-4a7c-9e11-4c3b58096a51" (UID: "7acc721f-ca1b-4a7c-9e11-4c3b58096a51"). InnerVolumeSpecName "kube-api-access-d7v5q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.012570 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "7acc721f-ca1b-4a7c-9e11-4c3b58096a51" (UID: "7acc721f-ca1b-4a7c-9e11-4c3b58096a51"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.012668 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ceph" (OuterVolumeSpecName: "ceph") pod "7acc721f-ca1b-4a7c-9e11-4c3b58096a51" (UID: "7acc721f-ca1b-4a7c-9e11-4c3b58096a51"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.032510 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "7acc721f-ca1b-4a7c-9e11-4c3b58096a51" (UID: "7acc721f-ca1b-4a7c-9e11-4c3b58096a51"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.034394 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-inventory" (OuterVolumeSpecName: "inventory") pod "7acc721f-ca1b-4a7c-9e11-4c3b58096a51" (UID: "7acc721f-ca1b-4a7c-9e11-4c3b58096a51"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.042912 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "7acc721f-ca1b-4a7c-9e11-4c3b58096a51" (UID: "7acc721f-ca1b-4a7c-9e11-4c3b58096a51"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.104505 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.104557 4787 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.104571 4787 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.104584 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.104601 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d7v5q\" (UniqueName: \"kubernetes.io/projected/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-kube-api-access-d7v5q\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.104614 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/7acc721f-ca1b-4a7c-9e11-4c3b58096a51-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.382493 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" event={"ID":"7acc721f-ca1b-4a7c-9e11-4c3b58096a51","Type":"ContainerDied","Data":"0cb11df18cb2d52c3cfb81018cef842aac345b525b2d2849177cd6f8a8f7908f"} Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.382551 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cb11df18cb2d52c3cfb81018cef842aac345b525b2d2849177cd6f8a8f7908f" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.382576 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.510961 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk"] Dec 03 18:19:36 crc kubenswrapper[4787]: E1203 18:19:36.511503 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" containerName="extract-content" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.511525 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" containerName="extract-content" Dec 03 18:19:36 crc kubenswrapper[4787]: E1203 18:19:36.511536 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerName="extract-utilities" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.511545 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerName="extract-utilities" Dec 03 18:19:36 crc kubenswrapper[4787]: E1203 18:19:36.511562 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" containerName="registry-server" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.511570 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" containerName="registry-server" Dec 03 18:19:36 crc kubenswrapper[4787]: E1203 18:19:36.511584 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" containerName="extract-utilities" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.511592 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" containerName="extract-utilities" Dec 03 18:19:36 crc kubenswrapper[4787]: E1203 18:19:36.511606 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerName="registry-server" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.511615 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerName="registry-server" Dec 03 18:19:36 crc kubenswrapper[4787]: E1203 18:19:36.511636 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerName="extract-content" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.511644 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerName="extract-content" Dec 03 18:19:36 crc kubenswrapper[4787]: E1203 18:19:36.511681 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7acc721f-ca1b-4a7c-9e11-4c3b58096a51" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.511689 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7acc721f-ca1b-4a7c-9e11-4c3b58096a51" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.511946 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7acc721f-ca1b-4a7c-9e11-4c3b58096a51" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.511970 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="212489b3-1bfd-44a0-8c31-358d3d20c5a1" containerName="registry-server" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.511998 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a1d2603-4d44-48ff-b89a-f0fec82572fd" containerName="registry-server" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.512898 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.516068 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.518231 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.518334 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.518613 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.518619 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.518730 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.518745 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.521151 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.521834 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.527743 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk"] Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.615322 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.615368 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.615424 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.615449 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.615503 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.615522 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vlltj\" (UniqueName: \"kubernetes.io/projected/4fea1a97-fa55-49e0-ab06-88ec531f4792-kube-api-access-vlltj\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.615553 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.615576 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.615627 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.615847 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.615882 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.717987 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.718103 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.718138 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.718214 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.718245 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vlltj\" (UniqueName: \"kubernetes.io/projected/4fea1a97-fa55-49e0-ab06-88ec531f4792-kube-api-access-vlltj\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.718340 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.718391 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.718509 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.718604 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.718689 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.718752 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.718887 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.719924 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.725267 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.726825 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.727830 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.729668 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.729942 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.730609 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.731632 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.732207 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.734753 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vlltj\" (UniqueName: \"kubernetes.io/projected/4fea1a97-fa55-49e0-ab06-88ec531f4792-kube-api-access-vlltj\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:36 crc kubenswrapper[4787]: I1203 18:19:36.828312 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:19:37 crc kubenswrapper[4787]: I1203 18:19:37.425666 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk"] Dec 03 18:19:38 crc kubenswrapper[4787]: I1203 18:19:38.429510 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" event={"ID":"4fea1a97-fa55-49e0-ab06-88ec531f4792","Type":"ContainerStarted","Data":"e7dc5ad82631eb4b4f6ee60af4ef2371a96a39fb45d1853bc5f191940338ecc7"} Dec 03 18:19:38 crc kubenswrapper[4787]: I1203 18:19:38.430167 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" event={"ID":"4fea1a97-fa55-49e0-ab06-88ec531f4792","Type":"ContainerStarted","Data":"57432be930f4fce51db6d62d5c921b0883915b4edac474868b3eeb5a1cb660c9"} Dec 03 18:19:38 crc kubenswrapper[4787]: I1203 18:19:38.456038 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" podStartSLOduration=1.92528078 podStartE2EDuration="2.455969931s" podCreationTimestamp="2025-12-03 18:19:36 +0000 UTC" firstStartedPulling="2025-12-03 18:19:37.424801509 +0000 UTC m=+3994.242272478" lastFinishedPulling="2025-12-03 18:19:37.95549065 +0000 UTC m=+3994.772961629" observedRunningTime="2025-12-03 18:19:38.45106816 +0000 UTC m=+3995.268539129" watchObservedRunningTime="2025-12-03 18:19:38.455969931 +0000 UTC m=+3995.273440900" Dec 03 18:19:46 crc kubenswrapper[4787]: I1203 18:19:46.766526 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:19:46 crc kubenswrapper[4787]: E1203 18:19:46.767325 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:19:57 crc kubenswrapper[4787]: I1203 18:19:57.766668 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:19:57 crc kubenswrapper[4787]: E1203 18:19:57.767456 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:20:10 crc kubenswrapper[4787]: I1203 18:20:10.766154 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:20:10 crc kubenswrapper[4787]: E1203 18:20:10.766971 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:20:21 crc kubenswrapper[4787]: I1203 18:20:21.766166 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:20:21 crc kubenswrapper[4787]: E1203 18:20:21.766936 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:20:36 crc kubenswrapper[4787]: I1203 18:20:36.766925 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:20:36 crc kubenswrapper[4787]: E1203 18:20:36.767919 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:20:51 crc kubenswrapper[4787]: I1203 18:20:51.766214 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:20:51 crc kubenswrapper[4787]: E1203 18:20:51.767208 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:21:05 crc kubenswrapper[4787]: I1203 18:21:05.767888 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:21:05 crc kubenswrapper[4787]: E1203 18:21:05.769008 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.576849 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q7x96"] Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.579725 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.604149 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q7x96"] Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.646205 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-utilities\") pod \"redhat-marketplace-q7x96\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.646365 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sv72z\" (UniqueName: \"kubernetes.io/projected/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-kube-api-access-sv72z\") pod \"redhat-marketplace-q7x96\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.646462 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-catalog-content\") pod \"redhat-marketplace-q7x96\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.748307 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sv72z\" (UniqueName: \"kubernetes.io/projected/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-kube-api-access-sv72z\") pod \"redhat-marketplace-q7x96\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.748405 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-catalog-content\") pod \"redhat-marketplace-q7x96\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.748465 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-utilities\") pod \"redhat-marketplace-q7x96\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.748929 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-utilities\") pod \"redhat-marketplace-q7x96\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.748960 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-catalog-content\") pod \"redhat-marketplace-q7x96\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.768743 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sv72z\" (UniqueName: \"kubernetes.io/projected/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-kube-api-access-sv72z\") pod \"redhat-marketplace-q7x96\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:07 crc kubenswrapper[4787]: I1203 18:21:07.910912 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:08 crc kubenswrapper[4787]: I1203 18:21:08.405945 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q7x96"] Dec 03 18:21:08 crc kubenswrapper[4787]: I1203 18:21:08.472495 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7x96" event={"ID":"b5730d6a-3e5f-48a3-883d-221bafaa3ac9","Type":"ContainerStarted","Data":"e53972b281d5f1fbc553cdc476c0d7d917da6722ea1cfe247ca99147a85f0142"} Dec 03 18:21:09 crc kubenswrapper[4787]: I1203 18:21:09.483910 4787 generic.go:334] "Generic (PLEG): container finished" podID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" containerID="54121b6726a4c7518871767bb2fe5a1c5cbf4c79fe04f15fa34d4b60382ca28e" exitCode=0 Dec 03 18:21:09 crc kubenswrapper[4787]: I1203 18:21:09.483994 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7x96" event={"ID":"b5730d6a-3e5f-48a3-883d-221bafaa3ac9","Type":"ContainerDied","Data":"54121b6726a4c7518871767bb2fe5a1c5cbf4c79fe04f15fa34d4b60382ca28e"} Dec 03 18:21:10 crc kubenswrapper[4787]: I1203 18:21:10.495961 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7x96" event={"ID":"b5730d6a-3e5f-48a3-883d-221bafaa3ac9","Type":"ContainerStarted","Data":"f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347"} Dec 03 18:21:11 crc kubenswrapper[4787]: I1203 18:21:11.515480 4787 generic.go:334] "Generic (PLEG): container finished" podID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" containerID="f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347" exitCode=0 Dec 03 18:21:11 crc kubenswrapper[4787]: I1203 18:21:11.515571 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7x96" event={"ID":"b5730d6a-3e5f-48a3-883d-221bafaa3ac9","Type":"ContainerDied","Data":"f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347"} Dec 03 18:21:13 crc kubenswrapper[4787]: I1203 18:21:13.540746 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7x96" event={"ID":"b5730d6a-3e5f-48a3-883d-221bafaa3ac9","Type":"ContainerStarted","Data":"1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26"} Dec 03 18:21:13 crc kubenswrapper[4787]: I1203 18:21:13.565201 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q7x96" podStartSLOduration=3.354203481 podStartE2EDuration="6.565181582s" podCreationTimestamp="2025-12-03 18:21:07 +0000 UTC" firstStartedPulling="2025-12-03 18:21:09.488150568 +0000 UTC m=+4086.305621527" lastFinishedPulling="2025-12-03 18:21:12.699128659 +0000 UTC m=+4089.516599628" observedRunningTime="2025-12-03 18:21:13.558419131 +0000 UTC m=+4090.375890100" watchObservedRunningTime="2025-12-03 18:21:13.565181582 +0000 UTC m=+4090.382652541" Dec 03 18:21:17 crc kubenswrapper[4787]: I1203 18:21:17.767186 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:21:17 crc kubenswrapper[4787]: E1203 18:21:17.768281 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:21:17 crc kubenswrapper[4787]: I1203 18:21:17.911502 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:17 crc kubenswrapper[4787]: I1203 18:21:17.911571 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:17 crc kubenswrapper[4787]: I1203 18:21:17.963895 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:18 crc kubenswrapper[4787]: I1203 18:21:18.639968 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:18 crc kubenswrapper[4787]: I1203 18:21:18.695762 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q7x96"] Dec 03 18:21:20 crc kubenswrapper[4787]: I1203 18:21:20.615192 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q7x96" podUID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" containerName="registry-server" containerID="cri-o://1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26" gracePeriod=2 Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.233661 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.373251 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-utilities\") pod \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.373459 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sv72z\" (UniqueName: \"kubernetes.io/projected/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-kube-api-access-sv72z\") pod \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.373613 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-catalog-content\") pod \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\" (UID: \"b5730d6a-3e5f-48a3-883d-221bafaa3ac9\") " Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.374480 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-utilities" (OuterVolumeSpecName: "utilities") pod "b5730d6a-3e5f-48a3-883d-221bafaa3ac9" (UID: "b5730d6a-3e5f-48a3-883d-221bafaa3ac9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.382763 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-kube-api-access-sv72z" (OuterVolumeSpecName: "kube-api-access-sv72z") pod "b5730d6a-3e5f-48a3-883d-221bafaa3ac9" (UID: "b5730d6a-3e5f-48a3-883d-221bafaa3ac9"). InnerVolumeSpecName "kube-api-access-sv72z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.403237 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b5730d6a-3e5f-48a3-883d-221bafaa3ac9" (UID: "b5730d6a-3e5f-48a3-883d-221bafaa3ac9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.475849 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sv72z\" (UniqueName: \"kubernetes.io/projected/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-kube-api-access-sv72z\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.475882 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.475891 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b5730d6a-3e5f-48a3-883d-221bafaa3ac9-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.632538 4787 generic.go:334] "Generic (PLEG): container finished" podID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" containerID="1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26" exitCode=0 Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.632607 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7x96" event={"ID":"b5730d6a-3e5f-48a3-883d-221bafaa3ac9","Type":"ContainerDied","Data":"1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26"} Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.632659 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q7x96" event={"ID":"b5730d6a-3e5f-48a3-883d-221bafaa3ac9","Type":"ContainerDied","Data":"e53972b281d5f1fbc553cdc476c0d7d917da6722ea1cfe247ca99147a85f0142"} Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.632659 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q7x96" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.632692 4787 scope.go:117] "RemoveContainer" containerID="1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.690384 4787 scope.go:117] "RemoveContainer" containerID="f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.717861 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q7x96"] Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.722270 4787 scope.go:117] "RemoveContainer" containerID="54121b6726a4c7518871767bb2fe5a1c5cbf4c79fe04f15fa34d4b60382ca28e" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.738953 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q7x96"] Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.788834 4787 scope.go:117] "RemoveContainer" containerID="1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26" Dec 03 18:21:21 crc kubenswrapper[4787]: E1203 18:21:21.789391 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26\": container with ID starting with 1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26 not found: ID does not exist" containerID="1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.790503 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26"} err="failed to get container status \"1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26\": rpc error: code = NotFound desc = could not find container \"1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26\": container with ID starting with 1f65bd9fd356b69c823ae7998a04c346c34e1946de90d92e4af0dda8ea93ac26 not found: ID does not exist" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.790539 4787 scope.go:117] "RemoveContainer" containerID="f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347" Dec 03 18:21:21 crc kubenswrapper[4787]: E1203 18:21:21.790896 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347\": container with ID starting with f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347 not found: ID does not exist" containerID="f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.790938 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347"} err="failed to get container status \"f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347\": rpc error: code = NotFound desc = could not find container \"f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347\": container with ID starting with f9ccfdf1019d954adc647d3f78037721b39a766cde4e6ad9191747a817b0a347 not found: ID does not exist" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.790966 4787 scope.go:117] "RemoveContainer" containerID="54121b6726a4c7518871767bb2fe5a1c5cbf4c79fe04f15fa34d4b60382ca28e" Dec 03 18:21:21 crc kubenswrapper[4787]: E1203 18:21:21.791312 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"54121b6726a4c7518871767bb2fe5a1c5cbf4c79fe04f15fa34d4b60382ca28e\": container with ID starting with 54121b6726a4c7518871767bb2fe5a1c5cbf4c79fe04f15fa34d4b60382ca28e not found: ID does not exist" containerID="54121b6726a4c7518871767bb2fe5a1c5cbf4c79fe04f15fa34d4b60382ca28e" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.791349 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"54121b6726a4c7518871767bb2fe5a1c5cbf4c79fe04f15fa34d4b60382ca28e"} err="failed to get container status \"54121b6726a4c7518871767bb2fe5a1c5cbf4c79fe04f15fa34d4b60382ca28e\": rpc error: code = NotFound desc = could not find container \"54121b6726a4c7518871767bb2fe5a1c5cbf4c79fe04f15fa34d4b60382ca28e\": container with ID starting with 54121b6726a4c7518871767bb2fe5a1c5cbf4c79fe04f15fa34d4b60382ca28e not found: ID does not exist" Dec 03 18:21:21 crc kubenswrapper[4787]: I1203 18:21:21.793639 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" path="/var/lib/kubelet/pods/b5730d6a-3e5f-48a3-883d-221bafaa3ac9/volumes" Dec 03 18:21:29 crc kubenswrapper[4787]: I1203 18:21:29.766631 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:21:29 crc kubenswrapper[4787]: E1203 18:21:29.767762 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:21:42 crc kubenswrapper[4787]: I1203 18:21:42.766794 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:21:42 crc kubenswrapper[4787]: E1203 18:21:42.767902 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:21:53 crc kubenswrapper[4787]: I1203 18:21:53.776293 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:21:53 crc kubenswrapper[4787]: E1203 18:21:53.777546 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:22:08 crc kubenswrapper[4787]: I1203 18:22:08.766878 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:22:08 crc kubenswrapper[4787]: E1203 18:22:08.768781 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.019578 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-g27cf"] Dec 03 18:22:09 crc kubenswrapper[4787]: E1203 18:22:09.020161 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" containerName="extract-utilities" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.020190 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" containerName="extract-utilities" Dec 03 18:22:09 crc kubenswrapper[4787]: E1203 18:22:09.020227 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" containerName="registry-server" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.020239 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" containerName="registry-server" Dec 03 18:22:09 crc kubenswrapper[4787]: E1203 18:22:09.020271 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" containerName="extract-content" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.020285 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" containerName="extract-content" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.038071 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5730d6a-3e5f-48a3-883d-221bafaa3ac9" containerName="registry-server" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.067394 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.072062 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g27cf"] Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.187010 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p45mh\" (UniqueName: \"kubernetes.io/projected/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-kube-api-access-p45mh\") pod \"redhat-operators-g27cf\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.187337 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-catalog-content\") pod \"redhat-operators-g27cf\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.187513 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-utilities\") pod \"redhat-operators-g27cf\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.290063 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p45mh\" (UniqueName: \"kubernetes.io/projected/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-kube-api-access-p45mh\") pod \"redhat-operators-g27cf\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.290192 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-catalog-content\") pod \"redhat-operators-g27cf\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.290359 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-utilities\") pod \"redhat-operators-g27cf\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.290800 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-catalog-content\") pod \"redhat-operators-g27cf\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.290839 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-utilities\") pod \"redhat-operators-g27cf\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.810197 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p45mh\" (UniqueName: \"kubernetes.io/projected/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-kube-api-access-p45mh\") pod \"redhat-operators-g27cf\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:09 crc kubenswrapper[4787]: I1203 18:22:09.999522 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:10 crc kubenswrapper[4787]: I1203 18:22:10.475105 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-g27cf"] Dec 03 18:22:11 crc kubenswrapper[4787]: I1203 18:22:11.269811 4787 generic.go:334] "Generic (PLEG): container finished" podID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" containerID="091b7955f7ef93f472b7f77ea2d288eb2fb1182cba02c23adf4530bac9c030f7" exitCode=0 Dec 03 18:22:11 crc kubenswrapper[4787]: I1203 18:22:11.269919 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g27cf" event={"ID":"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0","Type":"ContainerDied","Data":"091b7955f7ef93f472b7f77ea2d288eb2fb1182cba02c23adf4530bac9c030f7"} Dec 03 18:22:11 crc kubenswrapper[4787]: I1203 18:22:11.270294 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g27cf" event={"ID":"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0","Type":"ContainerStarted","Data":"8f217be6755dd1eaf240b159a1f2c9db75b0e405b2eb93d439b883aeae1bacc4"} Dec 03 18:22:13 crc kubenswrapper[4787]: I1203 18:22:13.296708 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g27cf" event={"ID":"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0","Type":"ContainerStarted","Data":"4ca042ea94385a153bb65fe8328234041b9aa0277afe94438f990d1d39af8129"} Dec 03 18:22:18 crc kubenswrapper[4787]: I1203 18:22:18.349269 4787 generic.go:334] "Generic (PLEG): container finished" podID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" containerID="4ca042ea94385a153bb65fe8328234041b9aa0277afe94438f990d1d39af8129" exitCode=0 Dec 03 18:22:18 crc kubenswrapper[4787]: I1203 18:22:18.349383 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g27cf" event={"ID":"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0","Type":"ContainerDied","Data":"4ca042ea94385a153bb65fe8328234041b9aa0277afe94438f990d1d39af8129"} Dec 03 18:22:23 crc kubenswrapper[4787]: I1203 18:22:23.420767 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g27cf" event={"ID":"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0","Type":"ContainerStarted","Data":"f81cb9b2950d04ff00c917ce3c2a71ae510f7d431ed70fb6a6b5fb6d5d807d18"} Dec 03 18:22:23 crc kubenswrapper[4787]: I1203 18:22:23.450509 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-g27cf" podStartSLOduration=4.354537102 podStartE2EDuration="15.450486444s" podCreationTimestamp="2025-12-03 18:22:08 +0000 UTC" firstStartedPulling="2025-12-03 18:22:11.271590107 +0000 UTC m=+4148.089061066" lastFinishedPulling="2025-12-03 18:22:22.367539429 +0000 UTC m=+4159.185010408" observedRunningTime="2025-12-03 18:22:23.445064999 +0000 UTC m=+4160.262536008" watchObservedRunningTime="2025-12-03 18:22:23.450486444 +0000 UTC m=+4160.267957423" Dec 03 18:22:23 crc kubenswrapper[4787]: I1203 18:22:23.780161 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:22:24 crc kubenswrapper[4787]: I1203 18:22:24.434939 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"b9b9fc7504675a616ee9e14cea9676e2597ec18a73f3b10eafe82561fc5ae2d1"} Dec 03 18:22:30 crc kubenswrapper[4787]: I1203 18:22:30.000014 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:30 crc kubenswrapper[4787]: I1203 18:22:30.000931 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:30 crc kubenswrapper[4787]: I1203 18:22:30.078807 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:30 crc kubenswrapper[4787]: I1203 18:22:30.583072 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:30 crc kubenswrapper[4787]: I1203 18:22:30.649522 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g27cf"] Dec 03 18:22:32 crc kubenswrapper[4787]: I1203 18:22:32.549598 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-g27cf" podUID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" containerName="registry-server" containerID="cri-o://f81cb9b2950d04ff00c917ce3c2a71ae510f7d431ed70fb6a6b5fb6d5d807d18" gracePeriod=2 Dec 03 18:22:33 crc kubenswrapper[4787]: I1203 18:22:33.565010 4787 generic.go:334] "Generic (PLEG): container finished" podID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" containerID="f81cb9b2950d04ff00c917ce3c2a71ae510f7d431ed70fb6a6b5fb6d5d807d18" exitCode=0 Dec 03 18:22:33 crc kubenswrapper[4787]: I1203 18:22:33.565110 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g27cf" event={"ID":"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0","Type":"ContainerDied","Data":"f81cb9b2950d04ff00c917ce3c2a71ae510f7d431ed70fb6a6b5fb6d5d807d18"} Dec 03 18:22:33 crc kubenswrapper[4787]: I1203 18:22:33.871956 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:33 crc kubenswrapper[4787]: I1203 18:22:33.992868 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-catalog-content\") pod \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " Dec 03 18:22:33 crc kubenswrapper[4787]: I1203 18:22:33.992914 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-utilities\") pod \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " Dec 03 18:22:33 crc kubenswrapper[4787]: I1203 18:22:33.993054 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p45mh\" (UniqueName: \"kubernetes.io/projected/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-kube-api-access-p45mh\") pod \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\" (UID: \"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0\") " Dec 03 18:22:33 crc kubenswrapper[4787]: I1203 18:22:33.993649 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-utilities" (OuterVolumeSpecName: "utilities") pod "c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" (UID: "c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.011844 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-kube-api-access-p45mh" (OuterVolumeSpecName: "kube-api-access-p45mh") pod "c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" (UID: "c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0"). InnerVolumeSpecName "kube-api-access-p45mh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.095924 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.095955 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p45mh\" (UniqueName: \"kubernetes.io/projected/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-kube-api-access-p45mh\") on node \"crc\" DevicePath \"\"" Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.100824 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" (UID: "c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.197310 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.586652 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-g27cf" event={"ID":"c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0","Type":"ContainerDied","Data":"8f217be6755dd1eaf240b159a1f2c9db75b0e405b2eb93d439b883aeae1bacc4"} Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.586725 4787 scope.go:117] "RemoveContainer" containerID="f81cb9b2950d04ff00c917ce3c2a71ae510f7d431ed70fb6a6b5fb6d5d807d18" Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.586747 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-g27cf" Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.640863 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-g27cf"] Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.646425 4787 scope.go:117] "RemoveContainer" containerID="4ca042ea94385a153bb65fe8328234041b9aa0277afe94438f990d1d39af8129" Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.668655 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-g27cf"] Dec 03 18:22:34 crc kubenswrapper[4787]: I1203 18:22:34.688542 4787 scope.go:117] "RemoveContainer" containerID="091b7955f7ef93f472b7f77ea2d288eb2fb1182cba02c23adf4530bac9c030f7" Dec 03 18:22:35 crc kubenswrapper[4787]: I1203 18:22:35.778106 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" path="/var/lib/kubelet/pods/c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0/volumes" Dec 03 18:23:51 crc kubenswrapper[4787]: I1203 18:23:51.488829 4787 generic.go:334] "Generic (PLEG): container finished" podID="4fea1a97-fa55-49e0-ab06-88ec531f4792" containerID="e7dc5ad82631eb4b4f6ee60af4ef2371a96a39fb45d1853bc5f191940338ecc7" exitCode=0 Dec 03 18:23:51 crc kubenswrapper[4787]: I1203 18:23:51.488964 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" event={"ID":"4fea1a97-fa55-49e0-ab06-88ec531f4792","Type":"ContainerDied","Data":"e7dc5ad82631eb4b4f6ee60af4ef2371a96a39fb45d1853bc5f191940338ecc7"} Dec 03 18:23:52 crc kubenswrapper[4787]: I1203 18:23:52.993282 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.064232 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-custom-ceph-combined-ca-bundle\") pod \"4fea1a97-fa55-49e0-ab06-88ec531f4792\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.064279 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-1\") pod \"4fea1a97-fa55-49e0-ab06-88ec531f4792\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.064347 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vlltj\" (UniqueName: \"kubernetes.io/projected/4fea1a97-fa55-49e0-ab06-88ec531f4792-kube-api-access-vlltj\") pod \"4fea1a97-fa55-49e0-ab06-88ec531f4792\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.064376 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph-nova-0\") pod \"4fea1a97-fa55-49e0-ab06-88ec531f4792\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.064400 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-inventory\") pod \"4fea1a97-fa55-49e0-ab06-88ec531f4792\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.064426 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph\") pod \"4fea1a97-fa55-49e0-ab06-88ec531f4792\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.064459 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-0\") pod \"4fea1a97-fa55-49e0-ab06-88ec531f4792\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.064555 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-1\") pod \"4fea1a97-fa55-49e0-ab06-88ec531f4792\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.064595 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-0\") pod \"4fea1a97-fa55-49e0-ab06-88ec531f4792\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.064685 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ssh-key\") pod \"4fea1a97-fa55-49e0-ab06-88ec531f4792\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.064724 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-extra-config-0\") pod \"4fea1a97-fa55-49e0-ab06-88ec531f4792\" (UID: \"4fea1a97-fa55-49e0-ab06-88ec531f4792\") " Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.083510 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "4fea1a97-fa55-49e0-ab06-88ec531f4792" (UID: "4fea1a97-fa55-49e0-ab06-88ec531f4792"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.083554 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fea1a97-fa55-49e0-ab06-88ec531f4792-kube-api-access-vlltj" (OuterVolumeSpecName: "kube-api-access-vlltj") pod "4fea1a97-fa55-49e0-ab06-88ec531f4792" (UID: "4fea1a97-fa55-49e0-ab06-88ec531f4792"). InnerVolumeSpecName "kube-api-access-vlltj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.084285 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph" (OuterVolumeSpecName: "ceph") pod "4fea1a97-fa55-49e0-ab06-88ec531f4792" (UID: "4fea1a97-fa55-49e0-ab06-88ec531f4792"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.110999 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "4fea1a97-fa55-49e0-ab06-88ec531f4792" (UID: "4fea1a97-fa55-49e0-ab06-88ec531f4792"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.124373 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "4fea1a97-fa55-49e0-ab06-88ec531f4792" (UID: "4fea1a97-fa55-49e0-ab06-88ec531f4792"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.126218 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4fea1a97-fa55-49e0-ab06-88ec531f4792" (UID: "4fea1a97-fa55-49e0-ab06-88ec531f4792"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.129159 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "4fea1a97-fa55-49e0-ab06-88ec531f4792" (UID: "4fea1a97-fa55-49e0-ab06-88ec531f4792"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.131669 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "4fea1a97-fa55-49e0-ab06-88ec531f4792" (UID: "4fea1a97-fa55-49e0-ab06-88ec531f4792"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.143642 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "4fea1a97-fa55-49e0-ab06-88ec531f4792" (UID: "4fea1a97-fa55-49e0-ab06-88ec531f4792"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.149955 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-inventory" (OuterVolumeSpecName: "inventory") pod "4fea1a97-fa55-49e0-ab06-88ec531f4792" (UID: "4fea1a97-fa55-49e0-ab06-88ec531f4792"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.150211 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "4fea1a97-fa55-49e0-ab06-88ec531f4792" (UID: "4fea1a97-fa55-49e0-ab06-88ec531f4792"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.167567 4787 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.167613 4787 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.167623 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.167632 4787 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.167642 4787 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.167654 4787 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.167664 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vlltj\" (UniqueName: \"kubernetes.io/projected/4fea1a97-fa55-49e0-ab06-88ec531f4792-kube-api-access-vlltj\") on node \"crc\" DevicePath \"\"" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.167673 4787 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.167682 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.167689 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.167697 4787 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/4fea1a97-fa55-49e0-ab06-88ec531f4792-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.511376 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" event={"ID":"4fea1a97-fa55-49e0-ab06-88ec531f4792","Type":"ContainerDied","Data":"57432be930f4fce51db6d62d5c921b0883915b4edac474868b3eeb5a1cb660c9"} Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.511654 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="57432be930f4fce51db6d62d5c921b0883915b4edac474868b3eeb5a1cb660c9" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.511470 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.639436 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf"] Dec 03 18:23:53 crc kubenswrapper[4787]: E1203 18:23:53.639908 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" containerName="extract-content" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.639925 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" containerName="extract-content" Dec 03 18:23:53 crc kubenswrapper[4787]: E1203 18:23:53.639962 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fea1a97-fa55-49e0-ab06-88ec531f4792" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.639969 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fea1a97-fa55-49e0-ab06-88ec531f4792" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 03 18:23:53 crc kubenswrapper[4787]: E1203 18:23:53.639979 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" containerName="registry-server" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.639985 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" containerName="registry-server" Dec 03 18:23:53 crc kubenswrapper[4787]: E1203 18:23:53.639996 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" containerName="extract-utilities" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.640003 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" containerName="extract-utilities" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.640229 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c427c26d-cc0e-49ca-9e7f-b1fe3ef5d6a0" containerName="registry-server" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.640253 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fea1a97-fa55-49e0-ab06-88ec531f4792" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.640964 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.643860 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.644100 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.644330 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.644945 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-compute-config-data" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.645101 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.645253 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.651391 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf"] Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.779117 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.779267 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kvdv\" (UniqueName: \"kubernetes.io/projected/93b51269-04f3-47d9-b357-9b274a2b3cfe-kube-api-access-6kvdv\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.779407 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.779593 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.779662 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.779917 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.780089 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.780211 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.882375 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.882475 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.882524 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.882598 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.882625 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kvdv\" (UniqueName: \"kubernetes.io/projected/93b51269-04f3-47d9-b357-9b274a2b3cfe-kube-api-access-6kvdv\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.882656 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.882758 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.882782 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.886806 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-2\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.887687 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-inventory\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.888217 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-telemetry-combined-ca-bundle\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.888513 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ssh-key\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.888528 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-1\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.889319 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceph\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.889988 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-0\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.915105 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kvdv\" (UniqueName: \"kubernetes.io/projected/93b51269-04f3-47d9-b357-9b274a2b3cfe-kube-api-access-6kvdv\") pod \"telemetry-edpm-deployment-openstack-edpm-ipam-c4phf\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:53 crc kubenswrapper[4787]: I1203 18:23:53.965047 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:23:54 crc kubenswrapper[4787]: I1203 18:23:54.499463 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf"] Dec 03 18:23:54 crc kubenswrapper[4787]: I1203 18:23:54.521861 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" event={"ID":"93b51269-04f3-47d9-b357-9b274a2b3cfe","Type":"ContainerStarted","Data":"7face68214b4096ec7b9bf69c2bbb126576b5b1b34ee56b3fa4859e5829e983b"} Dec 03 18:23:55 crc kubenswrapper[4787]: I1203 18:23:55.532062 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" event={"ID":"93b51269-04f3-47d9-b357-9b274a2b3cfe","Type":"ContainerStarted","Data":"76ff7679d08401556c67827c8b431786bdc440c885cf107acf9b6368ad69fa2a"} Dec 03 18:23:55 crc kubenswrapper[4787]: I1203 18:23:55.561739 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" podStartSLOduration=2.058514287 podStartE2EDuration="2.561711536s" podCreationTimestamp="2025-12-03 18:23:53 +0000 UTC" firstStartedPulling="2025-12-03 18:23:54.507044948 +0000 UTC m=+4251.324515907" lastFinishedPulling="2025-12-03 18:23:55.010242197 +0000 UTC m=+4251.827713156" observedRunningTime="2025-12-03 18:23:55.557380711 +0000 UTC m=+4252.374851720" watchObservedRunningTime="2025-12-03 18:23:55.561711536 +0000 UTC m=+4252.379182545" Dec 03 18:24:48 crc kubenswrapper[4787]: I1203 18:24:48.989756 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:24:48 crc kubenswrapper[4787]: I1203 18:24:48.990449 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:25:18 crc kubenswrapper[4787]: I1203 18:25:18.990008 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:25:18 crc kubenswrapper[4787]: I1203 18:25:18.990519 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:25:48 crc kubenswrapper[4787]: I1203 18:25:48.989665 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:25:48 crc kubenswrapper[4787]: I1203 18:25:48.990267 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:25:48 crc kubenswrapper[4787]: I1203 18:25:48.990314 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 18:25:48 crc kubenswrapper[4787]: I1203 18:25:48.991107 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b9b9fc7504675a616ee9e14cea9676e2597ec18a73f3b10eafe82561fc5ae2d1"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:25:48 crc kubenswrapper[4787]: I1203 18:25:48.991168 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://b9b9fc7504675a616ee9e14cea9676e2597ec18a73f3b10eafe82561fc5ae2d1" gracePeriod=600 Dec 03 18:25:49 crc kubenswrapper[4787]: I1203 18:25:49.856140 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="b9b9fc7504675a616ee9e14cea9676e2597ec18a73f3b10eafe82561fc5ae2d1" exitCode=0 Dec 03 18:25:49 crc kubenswrapper[4787]: I1203 18:25:49.856234 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"b9b9fc7504675a616ee9e14cea9676e2597ec18a73f3b10eafe82561fc5ae2d1"} Dec 03 18:25:49 crc kubenswrapper[4787]: I1203 18:25:49.856780 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3"} Dec 03 18:25:49 crc kubenswrapper[4787]: I1203 18:25:49.856818 4787 scope.go:117] "RemoveContainer" containerID="ee87a11d31db9a4e30ab742e4f0c182e1eb5facfdb107d767a3b0ebfa470b65c" Dec 03 18:27:10 crc kubenswrapper[4787]: I1203 18:27:10.929665 4787 generic.go:334] "Generic (PLEG): container finished" podID="93b51269-04f3-47d9-b357-9b274a2b3cfe" containerID="76ff7679d08401556c67827c8b431786bdc440c885cf107acf9b6368ad69fa2a" exitCode=0 Dec 03 18:27:10 crc kubenswrapper[4787]: I1203 18:27:10.929737 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" event={"ID":"93b51269-04f3-47d9-b357-9b274a2b3cfe","Type":"ContainerDied","Data":"76ff7679d08401556c67827c8b431786bdc440c885cf107acf9b6368ad69fa2a"} Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.429366 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.582910 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ssh-key\") pod \"93b51269-04f3-47d9-b357-9b274a2b3cfe\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.582996 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-1\") pod \"93b51269-04f3-47d9-b357-9b274a2b3cfe\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.583107 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-2\") pod \"93b51269-04f3-47d9-b357-9b274a2b3cfe\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.583167 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kvdv\" (UniqueName: \"kubernetes.io/projected/93b51269-04f3-47d9-b357-9b274a2b3cfe-kube-api-access-6kvdv\") pod \"93b51269-04f3-47d9-b357-9b274a2b3cfe\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.583195 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-telemetry-combined-ca-bundle\") pod \"93b51269-04f3-47d9-b357-9b274a2b3cfe\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.583221 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceph\") pod \"93b51269-04f3-47d9-b357-9b274a2b3cfe\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.583274 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-inventory\") pod \"93b51269-04f3-47d9-b357-9b274a2b3cfe\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.583327 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-0\") pod \"93b51269-04f3-47d9-b357-9b274a2b3cfe\" (UID: \"93b51269-04f3-47d9-b357-9b274a2b3cfe\") " Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.590047 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-telemetry-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-combined-ca-bundle") pod "93b51269-04f3-47d9-b357-9b274a2b3cfe" (UID: "93b51269-04f3-47d9-b357-9b274a2b3cfe"). InnerVolumeSpecName "telemetry-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.604227 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceph" (OuterVolumeSpecName: "ceph") pod "93b51269-04f3-47d9-b357-9b274a2b3cfe" (UID: "93b51269-04f3-47d9-b357-9b274a2b3cfe"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.619751 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93b51269-04f3-47d9-b357-9b274a2b3cfe-kube-api-access-6kvdv" (OuterVolumeSpecName: "kube-api-access-6kvdv") pod "93b51269-04f3-47d9-b357-9b274a2b3cfe" (UID: "93b51269-04f3-47d9-b357-9b274a2b3cfe"). InnerVolumeSpecName "kube-api-access-6kvdv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.625550 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-1" (OuterVolumeSpecName: "ceilometer-compute-config-data-1") pod "93b51269-04f3-47d9-b357-9b274a2b3cfe" (UID: "93b51269-04f3-47d9-b357-9b274a2b3cfe"). InnerVolumeSpecName "ceilometer-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.630097 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-inventory" (OuterVolumeSpecName: "inventory") pod "93b51269-04f3-47d9-b357-9b274a2b3cfe" (UID: "93b51269-04f3-47d9-b357-9b274a2b3cfe"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.635956 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "93b51269-04f3-47d9-b357-9b274a2b3cfe" (UID: "93b51269-04f3-47d9-b357-9b274a2b3cfe"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.648083 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-0" (OuterVolumeSpecName: "ceilometer-compute-config-data-0") pod "93b51269-04f3-47d9-b357-9b274a2b3cfe" (UID: "93b51269-04f3-47d9-b357-9b274a2b3cfe"). InnerVolumeSpecName "ceilometer-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.655241 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-2" (OuterVolumeSpecName: "ceilometer-compute-config-data-2") pod "93b51269-04f3-47d9-b357-9b274a2b3cfe" (UID: "93b51269-04f3-47d9-b357-9b274a2b3cfe"). InnerVolumeSpecName "ceilometer-compute-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.687217 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.687267 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.687287 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-2\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.687306 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kvdv\" (UniqueName: \"kubernetes.io/projected/93b51269-04f3-47d9-b357-9b274a2b3cfe-kube-api-access-6kvdv\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.687322 4787 reconciler_common.go:293] "Volume detached for volume \"telemetry-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-telemetry-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.687338 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.687353 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.687369 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/93b51269-04f3-47d9-b357-9b274a2b3cfe-ceilometer-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.962148 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" event={"ID":"93b51269-04f3-47d9-b357-9b274a2b3cfe","Type":"ContainerDied","Data":"7face68214b4096ec7b9bf69c2bbb126576b5b1b34ee56b3fa4859e5829e983b"} Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.962206 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7face68214b4096ec7b9bf69c2bbb126576b5b1b34ee56b3fa4859e5829e983b" Dec 03 18:27:12 crc kubenswrapper[4787]: I1203 18:27:12.962283 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-edpm-deployment-openstack-edpm-ipam-c4phf" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.075235 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq"] Dec 03 18:27:13 crc kubenswrapper[4787]: E1203 18:27:13.075848 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93b51269-04f3-47d9-b357-9b274a2b3cfe" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.075873 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="93b51269-04f3-47d9-b357-9b274a2b3cfe" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.076204 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="93b51269-04f3-47d9-b357-9b274a2b3cfe" containerName="telemetry-edpm-deployment-openstack-edpm-ipam" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.077224 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.079851 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.080225 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.081199 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.081216 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.081431 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-ipmi-config-data" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.081566 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.090168 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq"] Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.095053 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.095098 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.095124 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.095162 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jq5mq\" (UniqueName: \"kubernetes.io/projected/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-kube-api-access-jq5mq\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.095210 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.095282 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.095376 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.095409 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.197286 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.197367 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.197407 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.197432 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.197449 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.197477 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jq5mq\" (UniqueName: \"kubernetes.io/projected/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-kube-api-access-jq5mq\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.197509 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.197562 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.201721 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-inventory\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.202549 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-2\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.203454 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-0\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.204583 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-telemetry-power-monitoring-combined-ca-bundle\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.206772 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ssh-key\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.207367 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-1\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.207589 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceph\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.218364 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jq5mq\" (UniqueName: \"kubernetes.io/projected/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-kube-api-access-jq5mq\") pod \"telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:13 crc kubenswrapper[4787]: I1203 18:27:13.407743 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:27:14 crc kubenswrapper[4787]: I1203 18:27:14.002284 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq"] Dec 03 18:27:14 crc kubenswrapper[4787]: W1203 18:27:14.005774 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69ab81e9_ff7c_40f4_9f76_6f9e06e4caee.slice/crio-fd51fb3f66927bd1a33b63fcba1b36d57c0da15a657e00fe6e49d93baff43c34 WatchSource:0}: Error finding container fd51fb3f66927bd1a33b63fcba1b36d57c0da15a657e00fe6e49d93baff43c34: Status 404 returned error can't find the container with id fd51fb3f66927bd1a33b63fcba1b36d57c0da15a657e00fe6e49d93baff43c34 Dec 03 18:27:14 crc kubenswrapper[4787]: I1203 18:27:14.011957 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:27:14 crc kubenswrapper[4787]: I1203 18:27:14.983067 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" event={"ID":"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee","Type":"ContainerStarted","Data":"d8729340c8965603d3dba19f161304330ae062a0530ce2278ffca87fa3cff200"} Dec 03 18:27:14 crc kubenswrapper[4787]: I1203 18:27:14.983484 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" event={"ID":"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee","Type":"ContainerStarted","Data":"fd51fb3f66927bd1a33b63fcba1b36d57c0da15a657e00fe6e49d93baff43c34"} Dec 03 18:27:15 crc kubenswrapper[4787]: I1203 18:27:15.005918 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" podStartSLOduration=1.51215832 podStartE2EDuration="2.005887595s" podCreationTimestamp="2025-12-03 18:27:13 +0000 UTC" firstStartedPulling="2025-12-03 18:27:14.011696195 +0000 UTC m=+4450.829167164" lastFinishedPulling="2025-12-03 18:27:14.50542547 +0000 UTC m=+4451.322896439" observedRunningTime="2025-12-03 18:27:15.000777789 +0000 UTC m=+4451.818248758" watchObservedRunningTime="2025-12-03 18:27:15.005887595 +0000 UTC m=+4451.823358564" Dec 03 18:28:18 crc kubenswrapper[4787]: I1203 18:28:18.989884 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:28:18 crc kubenswrapper[4787]: I1203 18:28:18.990417 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:28:48 crc kubenswrapper[4787]: I1203 18:28:48.989904 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:28:48 crc kubenswrapper[4787]: I1203 18:28:48.990567 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:29:18 crc kubenswrapper[4787]: I1203 18:29:18.989668 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:29:18 crc kubenswrapper[4787]: I1203 18:29:18.990291 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:29:18 crc kubenswrapper[4787]: I1203 18:29:18.990354 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 18:29:18 crc kubenswrapper[4787]: I1203 18:29:18.991491 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:29:18 crc kubenswrapper[4787]: I1203 18:29:18.991595 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" gracePeriod=600 Dec 03 18:29:19 crc kubenswrapper[4787]: E1203 18:29:19.789400 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:29:20 crc kubenswrapper[4787]: I1203 18:29:20.378449 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" exitCode=0 Dec 03 18:29:20 crc kubenswrapper[4787]: I1203 18:29:20.378531 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3"} Dec 03 18:29:20 crc kubenswrapper[4787]: I1203 18:29:20.378772 4787 scope.go:117] "RemoveContainer" containerID="b9b9fc7504675a616ee9e14cea9676e2597ec18a73f3b10eafe82561fc5ae2d1" Dec 03 18:29:20 crc kubenswrapper[4787]: I1203 18:29:20.380315 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:29:20 crc kubenswrapper[4787]: E1203 18:29:20.380605 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:29:32 crc kubenswrapper[4787]: I1203 18:29:32.766141 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:29:32 crc kubenswrapper[4787]: E1203 18:29:32.767295 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:29:43 crc kubenswrapper[4787]: I1203 18:29:43.778057 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:29:43 crc kubenswrapper[4787]: E1203 18:29:43.778889 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:29:57 crc kubenswrapper[4787]: I1203 18:29:57.766392 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:29:57 crc kubenswrapper[4787]: E1203 18:29:57.767495 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.184860 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs"] Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.190491 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.194905 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.195200 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.200559 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs"] Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.289734 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/356d1186-2f36-4101-80d0-0d240814bb08-secret-volume\") pod \"collect-profiles-29413110-pdmfs\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.290074 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vhz7c\" (UniqueName: \"kubernetes.io/projected/356d1186-2f36-4101-80d0-0d240814bb08-kube-api-access-vhz7c\") pod \"collect-profiles-29413110-pdmfs\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.290176 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/356d1186-2f36-4101-80d0-0d240814bb08-config-volume\") pod \"collect-profiles-29413110-pdmfs\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.392368 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vhz7c\" (UniqueName: \"kubernetes.io/projected/356d1186-2f36-4101-80d0-0d240814bb08-kube-api-access-vhz7c\") pod \"collect-profiles-29413110-pdmfs\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.392434 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/356d1186-2f36-4101-80d0-0d240814bb08-config-volume\") pod \"collect-profiles-29413110-pdmfs\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.392481 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/356d1186-2f36-4101-80d0-0d240814bb08-secret-volume\") pod \"collect-profiles-29413110-pdmfs\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.393460 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/356d1186-2f36-4101-80d0-0d240814bb08-config-volume\") pod \"collect-profiles-29413110-pdmfs\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.397770 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/356d1186-2f36-4101-80d0-0d240814bb08-secret-volume\") pod \"collect-profiles-29413110-pdmfs\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.422636 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vhz7c\" (UniqueName: \"kubernetes.io/projected/356d1186-2f36-4101-80d0-0d240814bb08-kube-api-access-vhz7c\") pod \"collect-profiles-29413110-pdmfs\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.526969 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:00 crc kubenswrapper[4787]: I1203 18:30:00.984671 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs"] Dec 03 18:30:00 crc kubenswrapper[4787]: W1203 18:30:00.987401 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice/crio-98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6 WatchSource:0}: Error finding container 98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6: Status 404 returned error can't find the container with id 98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6 Dec 03 18:30:01 crc kubenswrapper[4787]: I1203 18:30:01.878626 4787 generic.go:334] "Generic (PLEG): container finished" podID="356d1186-2f36-4101-80d0-0d240814bb08" containerID="c760cab848a34b7412baaa1bbdc8c4673ad19efda3d72a618b3a8411de801ac0" exitCode=0 Dec 03 18:30:01 crc kubenswrapper[4787]: I1203 18:30:01.878687 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" event={"ID":"356d1186-2f36-4101-80d0-0d240814bb08","Type":"ContainerDied","Data":"c760cab848a34b7412baaa1bbdc8c4673ad19efda3d72a618b3a8411de801ac0"} Dec 03 18:30:01 crc kubenswrapper[4787]: I1203 18:30:01.880318 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" event={"ID":"356d1186-2f36-4101-80d0-0d240814bb08","Type":"ContainerStarted","Data":"98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6"} Dec 03 18:30:01 crc kubenswrapper[4787]: I1203 18:30:01.882045 4787 generic.go:334] "Generic (PLEG): container finished" podID="69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" containerID="d8729340c8965603d3dba19f161304330ae062a0530ce2278ffca87fa3cff200" exitCode=0 Dec 03 18:30:01 crc kubenswrapper[4787]: I1203 18:30:01.882108 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" event={"ID":"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee","Type":"ContainerDied","Data":"d8729340c8965603d3dba19f161304330ae062a0530ce2278ffca87fa3cff200"} Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.426397 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.431891 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.462814 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-1\") pod \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.462879 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/356d1186-2f36-4101-80d0-0d240814bb08-secret-volume\") pod \"356d1186-2f36-4101-80d0-0d240814bb08\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.462900 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/356d1186-2f36-4101-80d0-0d240814bb08-config-volume\") pod \"356d1186-2f36-4101-80d0-0d240814bb08\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.462967 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-telemetry-power-monitoring-combined-ca-bundle\") pod \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.463011 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vhz7c\" (UniqueName: \"kubernetes.io/projected/356d1186-2f36-4101-80d0-0d240814bb08-kube-api-access-vhz7c\") pod \"356d1186-2f36-4101-80d0-0d240814bb08\" (UID: \"356d1186-2f36-4101-80d0-0d240814bb08\") " Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.463070 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jq5mq\" (UniqueName: \"kubernetes.io/projected/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-kube-api-access-jq5mq\") pod \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.463102 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-inventory\") pod \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.463126 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceph\") pod \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.463155 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-2\") pod \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.463181 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ssh-key\") pod \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.463295 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-0\") pod \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\" (UID: \"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee\") " Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.463526 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/356d1186-2f36-4101-80d0-0d240814bb08-config-volume" (OuterVolumeSpecName: "config-volume") pod "356d1186-2f36-4101-80d0-0d240814bb08" (UID: "356d1186-2f36-4101-80d0-0d240814bb08"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.463789 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/356d1186-2f36-4101-80d0-0d240814bb08-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.474649 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/356d1186-2f36-4101-80d0-0d240814bb08-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "356d1186-2f36-4101-80d0-0d240814bb08" (UID: "356d1186-2f36-4101-80d0-0d240814bb08"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.474869 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/356d1186-2f36-4101-80d0-0d240814bb08-kube-api-access-vhz7c" (OuterVolumeSpecName: "kube-api-access-vhz7c") pod "356d1186-2f36-4101-80d0-0d240814bb08" (UID: "356d1186-2f36-4101-80d0-0d240814bb08"). InnerVolumeSpecName "kube-api-access-vhz7c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.476215 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-kube-api-access-jq5mq" (OuterVolumeSpecName: "kube-api-access-jq5mq") pod "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" (UID: "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee"). InnerVolumeSpecName "kube-api-access-jq5mq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.482917 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceph" (OuterVolumeSpecName: "ceph") pod "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" (UID: "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.483244 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-telemetry-power-monitoring-combined-ca-bundle" (OuterVolumeSpecName: "telemetry-power-monitoring-combined-ca-bundle") pod "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" (UID: "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee"). InnerVolumeSpecName "telemetry-power-monitoring-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.506725 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-inventory" (OuterVolumeSpecName: "inventory") pod "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" (UID: "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.512539 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" (UID: "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.518803 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-0" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-0") pod "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" (UID: "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee"). InnerVolumeSpecName "ceilometer-ipmi-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.521492 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-1" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-1") pod "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" (UID: "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee"). InnerVolumeSpecName "ceilometer-ipmi-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.523279 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-2" (OuterVolumeSpecName: "ceilometer-ipmi-config-data-2") pod "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" (UID: "69ab81e9-ff7c-40f4-9f76-6f9e06e4caee"). InnerVolumeSpecName "ceilometer-ipmi-config-data-2". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.568133 4787 reconciler_common.go:293] "Volume detached for volume \"telemetry-power-monitoring-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-telemetry-power-monitoring-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.568190 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vhz7c\" (UniqueName: \"kubernetes.io/projected/356d1186-2f36-4101-80d0-0d240814bb08-kube-api-access-vhz7c\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.568204 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jq5mq\" (UniqueName: \"kubernetes.io/projected/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-kube-api-access-jq5mq\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.568216 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.568228 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.568241 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-2\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-2\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.568252 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.568266 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-0\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.568277 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-ipmi-config-data-1\" (UniqueName: \"kubernetes.io/secret/69ab81e9-ff7c-40f4-9f76-6f9e06e4caee-ceilometer-ipmi-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.568290 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/356d1186-2f36-4101-80d0-0d240814bb08-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.908367 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.908378 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs" event={"ID":"356d1186-2f36-4101-80d0-0d240814bb08","Type":"ContainerDied","Data":"98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6"} Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.908991 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.910936 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" event={"ID":"69ab81e9-ff7c-40f4-9f76-6f9e06e4caee","Type":"ContainerDied","Data":"fd51fb3f66927bd1a33b63fcba1b36d57c0da15a657e00fe6e49d93baff43c34"} Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.911003 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd51fb3f66927bd1a33b63fcba1b36d57c0da15a657e00fe6e49d93baff43c34" Dec 03 18:30:03 crc kubenswrapper[4787]: I1203 18:30:03.911128 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.045104 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb"] Dec 03 18:30:04 crc kubenswrapper[4787]: E1203 18:30:04.045494 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.045512 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 18:30:04 crc kubenswrapper[4787]: E1203 18:30:04.045534 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="356d1186-2f36-4101-80d0-0d240814bb08" containerName="collect-profiles" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.045541 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="356d1186-2f36-4101-80d0-0d240814bb08" containerName="collect-profiles" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.045756 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="356d1186-2f36-4101-80d0-0d240814bb08" containerName="collect-profiles" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.045788 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="69ab81e9-ff7c-40f4-9f76-6f9e06e4caee" containerName="telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.046546 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.049007 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.049353 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.049503 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.051354 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.051561 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"logging-compute-config-data" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.051650 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-m2v8z" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.062347 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb"] Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.077765 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.077863 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.077938 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.077999 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzx6v\" (UniqueName: \"kubernetes.io/projected/4f182e93-6f85-4e36-978f-50a273b371f0-kube-api-access-lzx6v\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.078083 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.078219 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.180729 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.180792 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.180852 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.180918 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.180953 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzx6v\" (UniqueName: \"kubernetes.io/projected/4f182e93-6f85-4e36-978f-50a273b371f0-kube-api-access-lzx6v\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.181360 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.187730 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ssh-key\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.196567 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-0\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.197357 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ceph\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.197628 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-inventory\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.202114 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-1\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.219728 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzx6v\" (UniqueName: \"kubernetes.io/projected/4f182e93-6f85-4e36-978f-50a273b371f0-kube-api-access-lzx6v\") pod \"logging-edpm-deployment-openstack-edpm-ipam-p8gvb\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.367955 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.516346 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr"] Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.525899 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413065-8gbhr"] Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.892395 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb"] Dec 03 18:30:04 crc kubenswrapper[4787]: W1203 18:30:04.899381 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f182e93_6f85_4e36_978f_50a273b371f0.slice/crio-e92b419c771f69ab332f2a4fe42f36dfc2e7082fbec0780aceb6dc25fc9efea6 WatchSource:0}: Error finding container e92b419c771f69ab332f2a4fe42f36dfc2e7082fbec0780aceb6dc25fc9efea6: Status 404 returned error can't find the container with id e92b419c771f69ab332f2a4fe42f36dfc2e7082fbec0780aceb6dc25fc9efea6 Dec 03 18:30:04 crc kubenswrapper[4787]: I1203 18:30:04.923491 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" event={"ID":"4f182e93-6f85-4e36-978f-50a273b371f0","Type":"ContainerStarted","Data":"e92b419c771f69ab332f2a4fe42f36dfc2e7082fbec0780aceb6dc25fc9efea6"} Dec 03 18:30:05 crc kubenswrapper[4787]: I1203 18:30:05.783514 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="512a131d-22af-40e4-b9ca-a8c04ff844de" path="/var/lib/kubelet/pods/512a131d-22af-40e4-b9ca-a8c04ff844de/volumes" Dec 03 18:30:05 crc kubenswrapper[4787]: I1203 18:30:05.937512 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" event={"ID":"4f182e93-6f85-4e36-978f-50a273b371f0","Type":"ContainerStarted","Data":"5314a5cf3c0f5d45ea68f28306de3189b17c832a77ab5830cf7fa0617dbd15be"} Dec 03 18:30:05 crc kubenswrapper[4787]: I1203 18:30:05.958905 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" podStartSLOduration=1.350921024 podStartE2EDuration="1.958888279s" podCreationTimestamp="2025-12-03 18:30:04 +0000 UTC" firstStartedPulling="2025-12-03 18:30:04.90398361 +0000 UTC m=+4621.721454569" lastFinishedPulling="2025-12-03 18:30:05.511950825 +0000 UTC m=+4622.329421824" observedRunningTime="2025-12-03 18:30:05.953696341 +0000 UTC m=+4622.771167330" watchObservedRunningTime="2025-12-03 18:30:05.958888279 +0000 UTC m=+4622.776359238" Dec 03 18:30:08 crc kubenswrapper[4787]: E1203 18:30:08.764809 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice/crio-98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69ab81e9_ff7c_40f4_9f76_6f9e06e4caee.slice\": RecentStats: unable to find data in memory cache]" Dec 03 18:30:10 crc kubenswrapper[4787]: I1203 18:30:10.766539 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:30:10 crc kubenswrapper[4787]: E1203 18:30:10.767106 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.751212 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-r2th2"] Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.754656 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.764931 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r2th2"] Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.860667 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-utilities\") pod \"certified-operators-r2th2\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.860735 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzxfv\" (UniqueName: \"kubernetes.io/projected/1ab4dda6-f04f-400a-9bad-da640b07f559-kube-api-access-fzxfv\") pod \"certified-operators-r2th2\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.861343 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-catalog-content\") pod \"certified-operators-r2th2\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.963619 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-utilities\") pod \"certified-operators-r2th2\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.963726 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzxfv\" (UniqueName: \"kubernetes.io/projected/1ab4dda6-f04f-400a-9bad-da640b07f559-kube-api-access-fzxfv\") pod \"certified-operators-r2th2\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.963807 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-catalog-content\") pod \"certified-operators-r2th2\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.965874 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-utilities\") pod \"certified-operators-r2th2\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.966216 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-catalog-content\") pod \"certified-operators-r2th2\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:12 crc kubenswrapper[4787]: I1203 18:30:12.984101 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzxfv\" (UniqueName: \"kubernetes.io/projected/1ab4dda6-f04f-400a-9bad-da640b07f559-kube-api-access-fzxfv\") pod \"certified-operators-r2th2\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:13 crc kubenswrapper[4787]: I1203 18:30:13.084099 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:13 crc kubenswrapper[4787]: I1203 18:30:13.764198 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-r2th2"] Dec 03 18:30:14 crc kubenswrapper[4787]: I1203 18:30:14.036550 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2th2" event={"ID":"1ab4dda6-f04f-400a-9bad-da640b07f559","Type":"ContainerStarted","Data":"151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe"} Dec 03 18:30:14 crc kubenswrapper[4787]: I1203 18:30:14.036596 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2th2" event={"ID":"1ab4dda6-f04f-400a-9bad-da640b07f559","Type":"ContainerStarted","Data":"50fc0c969c9df4426c51ea9dcac741142a33ae31e790e8b638642c7a083f1a8b"} Dec 03 18:30:15 crc kubenswrapper[4787]: I1203 18:30:15.048740 4787 generic.go:334] "Generic (PLEG): container finished" podID="1ab4dda6-f04f-400a-9bad-da640b07f559" containerID="151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe" exitCode=0 Dec 03 18:30:15 crc kubenswrapper[4787]: I1203 18:30:15.048815 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2th2" event={"ID":"1ab4dda6-f04f-400a-9bad-da640b07f559","Type":"ContainerDied","Data":"151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe"} Dec 03 18:30:15 crc kubenswrapper[4787]: I1203 18:30:15.049232 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2th2" event={"ID":"1ab4dda6-f04f-400a-9bad-da640b07f559","Type":"ContainerStarted","Data":"d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779"} Dec 03 18:30:16 crc kubenswrapper[4787]: I1203 18:30:16.062228 4787 generic.go:334] "Generic (PLEG): container finished" podID="1ab4dda6-f04f-400a-9bad-da640b07f559" containerID="d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779" exitCode=0 Dec 03 18:30:16 crc kubenswrapper[4787]: I1203 18:30:16.062356 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2th2" event={"ID":"1ab4dda6-f04f-400a-9bad-da640b07f559","Type":"ContainerDied","Data":"d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779"} Dec 03 18:30:18 crc kubenswrapper[4787]: I1203 18:30:18.355676 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2th2" event={"ID":"1ab4dda6-f04f-400a-9bad-da640b07f559","Type":"ContainerStarted","Data":"49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087"} Dec 03 18:30:18 crc kubenswrapper[4787]: I1203 18:30:18.381902 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-r2th2" podStartSLOduration=3.062331182 podStartE2EDuration="6.381885793s" podCreationTimestamp="2025-12-03 18:30:12 +0000 UTC" firstStartedPulling="2025-12-03 18:30:14.038240177 +0000 UTC m=+4630.855711136" lastFinishedPulling="2025-12-03 18:30:17.357794778 +0000 UTC m=+4634.175265747" observedRunningTime="2025-12-03 18:30:18.381538114 +0000 UTC m=+4635.199009073" watchObservedRunningTime="2025-12-03 18:30:18.381885793 +0000 UTC m=+4635.199356752" Dec 03 18:30:19 crc kubenswrapper[4787]: E1203 18:30:19.080150 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69ab81e9_ff7c_40f4_9f76_6f9e06e4caee.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice/crio-98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice\": RecentStats: unable to find data in memory cache]" Dec 03 18:30:21 crc kubenswrapper[4787]: I1203 18:30:21.766660 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:30:21 crc kubenswrapper[4787]: E1203 18:30:21.767585 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:30:22 crc kubenswrapper[4787]: I1203 18:30:22.399283 4787 generic.go:334] "Generic (PLEG): container finished" podID="4f182e93-6f85-4e36-978f-50a273b371f0" containerID="5314a5cf3c0f5d45ea68f28306de3189b17c832a77ab5830cf7fa0617dbd15be" exitCode=0 Dec 03 18:30:22 crc kubenswrapper[4787]: I1203 18:30:22.399366 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" event={"ID":"4f182e93-6f85-4e36-978f-50a273b371f0","Type":"ContainerDied","Data":"5314a5cf3c0f5d45ea68f28306de3189b17c832a77ab5830cf7fa0617dbd15be"} Dec 03 18:30:23 crc kubenswrapper[4787]: I1203 18:30:23.084974 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:23 crc kubenswrapper[4787]: I1203 18:30:23.085346 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:23 crc kubenswrapper[4787]: I1203 18:30:23.215587 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:23 crc kubenswrapper[4787]: I1203 18:30:23.459763 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:23 crc kubenswrapper[4787]: I1203 18:30:23.506362 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r2th2"] Dec 03 18:30:23 crc kubenswrapper[4787]: I1203 18:30:23.887980 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.015599 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-1\") pod \"4f182e93-6f85-4e36-978f-50a273b371f0\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.015823 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-inventory\") pod \"4f182e93-6f85-4e36-978f-50a273b371f0\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.016112 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ssh-key\") pod \"4f182e93-6f85-4e36-978f-50a273b371f0\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.016154 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ceph\") pod \"4f182e93-6f85-4e36-978f-50a273b371f0\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.016187 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-0\") pod \"4f182e93-6f85-4e36-978f-50a273b371f0\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.016251 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzx6v\" (UniqueName: \"kubernetes.io/projected/4f182e93-6f85-4e36-978f-50a273b371f0-kube-api-access-lzx6v\") pod \"4f182e93-6f85-4e36-978f-50a273b371f0\" (UID: \"4f182e93-6f85-4e36-978f-50a273b371f0\") " Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.022503 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ceph" (OuterVolumeSpecName: "ceph") pod "4f182e93-6f85-4e36-978f-50a273b371f0" (UID: "4f182e93-6f85-4e36-978f-50a273b371f0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.024507 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f182e93-6f85-4e36-978f-50a273b371f0-kube-api-access-lzx6v" (OuterVolumeSpecName: "kube-api-access-lzx6v") pod "4f182e93-6f85-4e36-978f-50a273b371f0" (UID: "4f182e93-6f85-4e36-978f-50a273b371f0"). InnerVolumeSpecName "kube-api-access-lzx6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.047377 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-0" (OuterVolumeSpecName: "logging-compute-config-data-0") pod "4f182e93-6f85-4e36-978f-50a273b371f0" (UID: "4f182e93-6f85-4e36-978f-50a273b371f0"). InnerVolumeSpecName "logging-compute-config-data-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.049333 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-1" (OuterVolumeSpecName: "logging-compute-config-data-1") pod "4f182e93-6f85-4e36-978f-50a273b371f0" (UID: "4f182e93-6f85-4e36-978f-50a273b371f0"). InnerVolumeSpecName "logging-compute-config-data-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.056714 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-inventory" (OuterVolumeSpecName: "inventory") pod "4f182e93-6f85-4e36-978f-50a273b371f0" (UID: "4f182e93-6f85-4e36-978f-50a273b371f0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.065518 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4f182e93-6f85-4e36-978f-50a273b371f0" (UID: "4f182e93-6f85-4e36-978f-50a273b371f0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.119889 4787 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-1\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-1\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.120283 4787 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-inventory\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.120305 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.120321 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.120338 4787 reconciler_common.go:293] "Volume detached for volume \"logging-compute-config-data-0\" (UniqueName: \"kubernetes.io/secret/4f182e93-6f85-4e36-978f-50a273b371f0-logging-compute-config-data-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.120359 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzx6v\" (UniqueName: \"kubernetes.io/projected/4f182e93-6f85-4e36-978f-50a273b371f0-kube-api-access-lzx6v\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.422712 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" event={"ID":"4f182e93-6f85-4e36-978f-50a273b371f0","Type":"ContainerDied","Data":"e92b419c771f69ab332f2a4fe42f36dfc2e7082fbec0780aceb6dc25fc9efea6"} Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.422758 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/logging-edpm-deployment-openstack-edpm-ipam-p8gvb" Dec 03 18:30:24 crc kubenswrapper[4787]: I1203 18:30:24.422767 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e92b419c771f69ab332f2a4fe42f36dfc2e7082fbec0780aceb6dc25fc9efea6" Dec 03 18:30:25 crc kubenswrapper[4787]: I1203 18:30:25.431275 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-r2th2" podUID="1ab4dda6-f04f-400a-9bad-da640b07f559" containerName="registry-server" containerID="cri-o://49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087" gracePeriod=2 Dec 03 18:30:25 crc kubenswrapper[4787]: I1203 18:30:25.981810 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.162048 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-catalog-content\") pod \"1ab4dda6-f04f-400a-9bad-da640b07f559\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.162690 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzxfv\" (UniqueName: \"kubernetes.io/projected/1ab4dda6-f04f-400a-9bad-da640b07f559-kube-api-access-fzxfv\") pod \"1ab4dda6-f04f-400a-9bad-da640b07f559\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.162861 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-utilities\") pod \"1ab4dda6-f04f-400a-9bad-da640b07f559\" (UID: \"1ab4dda6-f04f-400a-9bad-da640b07f559\") " Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.163522 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-utilities" (OuterVolumeSpecName: "utilities") pod "1ab4dda6-f04f-400a-9bad-da640b07f559" (UID: "1ab4dda6-f04f-400a-9bad-da640b07f559"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.169575 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ab4dda6-f04f-400a-9bad-da640b07f559-kube-api-access-fzxfv" (OuterVolumeSpecName: "kube-api-access-fzxfv") pod "1ab4dda6-f04f-400a-9bad-da640b07f559" (UID: "1ab4dda6-f04f-400a-9bad-da640b07f559"). InnerVolumeSpecName "kube-api-access-fzxfv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.225559 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1ab4dda6-f04f-400a-9bad-da640b07f559" (UID: "1ab4dda6-f04f-400a-9bad-da640b07f559"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.266708 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.266758 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzxfv\" (UniqueName: \"kubernetes.io/projected/1ab4dda6-f04f-400a-9bad-da640b07f559-kube-api-access-fzxfv\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.266774 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1ab4dda6-f04f-400a-9bad-da640b07f559-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.444001 4787 generic.go:334] "Generic (PLEG): container finished" podID="1ab4dda6-f04f-400a-9bad-da640b07f559" containerID="49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087" exitCode=0 Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.444082 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-r2th2" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.444103 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2th2" event={"ID":"1ab4dda6-f04f-400a-9bad-da640b07f559","Type":"ContainerDied","Data":"49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087"} Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.445220 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-r2th2" event={"ID":"1ab4dda6-f04f-400a-9bad-da640b07f559","Type":"ContainerDied","Data":"50fc0c969c9df4426c51ea9dcac741142a33ae31e790e8b638642c7a083f1a8b"} Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.445243 4787 scope.go:117] "RemoveContainer" containerID="49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.472374 4787 scope.go:117] "RemoveContainer" containerID="d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.519297 4787 scope.go:117] "RemoveContainer" containerID="151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.527091 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-r2th2"] Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.536998 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-r2th2"] Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.592994 4787 scope.go:117] "RemoveContainer" containerID="49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087" Dec 03 18:30:26 crc kubenswrapper[4787]: E1203 18:30:26.593429 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087\": container with ID starting with 49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087 not found: ID does not exist" containerID="49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.593474 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087"} err="failed to get container status \"49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087\": rpc error: code = NotFound desc = could not find container \"49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087\": container with ID starting with 49c574ce57c78eee1f5fae87c030c291f982b41727ecda363e1a8ac864d9a087 not found: ID does not exist" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.593500 4787 scope.go:117] "RemoveContainer" containerID="d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779" Dec 03 18:30:26 crc kubenswrapper[4787]: E1203 18:30:26.593833 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779\": container with ID starting with d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779 not found: ID does not exist" containerID="d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.593900 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779"} err="failed to get container status \"d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779\": rpc error: code = NotFound desc = could not find container \"d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779\": container with ID starting with d3835ecf9c27e5d3215b7986206a834fce529ed68f4c1195afe5bacf390eb779 not found: ID does not exist" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.593926 4787 scope.go:117] "RemoveContainer" containerID="151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe" Dec 03 18:30:26 crc kubenswrapper[4787]: E1203 18:30:26.594316 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe\": container with ID starting with 151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe not found: ID does not exist" containerID="151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe" Dec 03 18:30:26 crc kubenswrapper[4787]: I1203 18:30:26.594345 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe"} err="failed to get container status \"151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe\": rpc error: code = NotFound desc = could not find container \"151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe\": container with ID starting with 151a4ab94ce3cdf133840a559288b146a1fa4db0cd3a9f200bb3947dbd1f04fe not found: ID does not exist" Dec 03 18:30:27 crc kubenswrapper[4787]: I1203 18:30:27.332628 4787 scope.go:117] "RemoveContainer" containerID="112d87f048f7a3e087df6163dd9fca2bcca1a218762dc78a9eadd669e6d3af1c" Dec 03 18:30:27 crc kubenswrapper[4787]: I1203 18:30:27.781504 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ab4dda6-f04f-400a-9bad-da640b07f559" path="/var/lib/kubelet/pods/1ab4dda6-f04f-400a-9bad-da640b07f559/volumes" Dec 03 18:30:29 crc kubenswrapper[4787]: E1203 18:30:29.394881 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice/crio-98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69ab81e9_ff7c_40f4_9f76_6f9e06e4caee.slice\": RecentStats: unable to find data in memory cache]" Dec 03 18:30:36 crc kubenswrapper[4787]: I1203 18:30:36.766566 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:30:36 crc kubenswrapper[4787]: E1203 18:30:36.767270 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:30:39 crc kubenswrapper[4787]: E1203 18:30:39.683682 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice/crio-98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69ab81e9_ff7c_40f4_9f76_6f9e06e4caee.slice\": RecentStats: unable to find data in memory cache]" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.590599 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 03 18:30:40 crc kubenswrapper[4787]: E1203 18:30:40.591479 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ab4dda6-f04f-400a-9bad-da640b07f559" containerName="extract-utilities" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.591507 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ab4dda6-f04f-400a-9bad-da640b07f559" containerName="extract-utilities" Dec 03 18:30:40 crc kubenswrapper[4787]: E1203 18:30:40.591532 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ab4dda6-f04f-400a-9bad-da640b07f559" containerName="extract-content" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.591540 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ab4dda6-f04f-400a-9bad-da640b07f559" containerName="extract-content" Dec 03 18:30:40 crc kubenswrapper[4787]: E1203 18:30:40.591566 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ab4dda6-f04f-400a-9bad-da640b07f559" containerName="registry-server" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.591573 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ab4dda6-f04f-400a-9bad-da640b07f559" containerName="registry-server" Dec 03 18:30:40 crc kubenswrapper[4787]: E1203 18:30:40.591601 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f182e93-6f85-4e36-978f-50a273b371f0" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.591609 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f182e93-6f85-4e36-978f-50a273b371f0" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.591897 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f182e93-6f85-4e36-978f-50a273b371f0" containerName="logging-edpm-deployment-openstack-edpm-ipam" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.591936 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ab4dda6-f04f-400a-9bad-da640b07f559" containerName="registry-server" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.593366 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.595134 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.595284 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.606825 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693160 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693250 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-dev\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693275 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693303 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-sys\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693325 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693353 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693376 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693390 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693422 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693441 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693471 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693487 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzrcm\" (UniqueName: \"kubernetes.io/projected/0f2fd27e-97a7-4019-98fd-1ed092285098-kube-api-access-rzrcm\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693502 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-run\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693526 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0f2fd27e-97a7-4019-98fd-1ed092285098-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693549 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.693571 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.703756 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.706046 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.708388 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.725468 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.796778 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-etc-nvme\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.796853 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-dev\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.796884 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-config-data-custom\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.796926 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.796977 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.796982 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-sys\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.797094 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.797129 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.797160 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-run\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.797203 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-dev\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.797241 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.797262 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-scripts\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.797281 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.797329 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-sys\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.797357 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.797443 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/77e12136-eaca-4c6a-9346-da3325061b57-ceph\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.797444 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.798291 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-sys\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.798307 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.798390 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-dev\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799087 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799117 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799162 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-config-data\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799198 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799243 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799288 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799314 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799323 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzrcm\" (UniqueName: \"kubernetes.io/projected/0f2fd27e-97a7-4019-98fd-1ed092285098-kube-api-access-rzrcm\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799375 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-run\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799418 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799447 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0f2fd27e-97a7-4019-98fd-1ed092285098-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799469 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799485 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799493 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-glwvd\" (UniqueName: \"kubernetes.io/projected/77e12136-eaca-4c6a-9346-da3325061b57-kube-api-access-glwvd\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799600 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799662 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799673 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-run\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799705 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-lib-modules\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.799837 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.800055 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.800187 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.800247 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0f2fd27e-97a7-4019-98fd-1ed092285098-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.804913 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.806171 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0f2fd27e-97a7-4019-98fd-1ed092285098-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.806203 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.812243 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.827402 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f2fd27e-97a7-4019-98fd-1ed092285098-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.827817 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzrcm\" (UniqueName: \"kubernetes.io/projected/0f2fd27e-97a7-4019-98fd-1ed092285098-kube-api-access-rzrcm\") pod \"cinder-volume-volume1-0\" (UID: \"0f2fd27e-97a7-4019-98fd-1ed092285098\") " pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902329 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/77e12136-eaca-4c6a-9346-da3325061b57-ceph\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902402 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-config-data\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902469 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902495 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902518 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-glwvd\" (UniqueName: \"kubernetes.io/projected/77e12136-eaca-4c6a-9346-da3325061b57-kube-api-access-glwvd\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902555 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-lib-modules\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902576 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902614 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-etc-nvme\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902650 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-dev\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902673 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-config-data-custom\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902710 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-sys\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902743 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902770 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902798 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-run\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902842 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-scripts\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902864 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902882 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-lib-modules\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902899 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902929 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-sys\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902937 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.902998 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-run\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.903034 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.903032 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.903052 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-dev\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.903069 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-etc-nvme\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.903084 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/77e12136-eaca-4c6a-9346-da3325061b57-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.905968 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-config-data-custom\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.906494 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-config-data\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.907537 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-scripts\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.907545 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/77e12136-eaca-4c6a-9346-da3325061b57-ceph\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.907749 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/77e12136-eaca-4c6a-9346-da3325061b57-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.910896 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:40 crc kubenswrapper[4787]: I1203 18:30:40.923542 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-glwvd\" (UniqueName: \"kubernetes.io/projected/77e12136-eaca-4c6a-9346-da3325061b57-kube-api-access-glwvd\") pod \"cinder-backup-0\" (UID: \"77e12136-eaca-4c6a-9346-da3325061b57\") " pod="openstack/cinder-backup-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.027781 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.513274 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-2dp4k"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.514904 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2dp4k" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.535551 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-operator-scripts\") pod \"manila-db-create-2dp4k\" (UID: \"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de\") " pod="openstack/manila-db-create-2dp4k" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.535603 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dw9wt\" (UniqueName: \"kubernetes.io/projected/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-kube-api-access-dw9wt\") pod \"manila-db-create-2dp4k\" (UID: \"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de\") " pod="openstack/manila-db-create-2dp4k" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.542084 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-2dp4k"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.580475 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.582332 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.589069 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-8g2xg" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.589239 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.589949 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.590174 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.642948 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-operator-scripts\") pod \"manila-db-create-2dp4k\" (UID: \"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de\") " pod="openstack/manila-db-create-2dp4k" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.643004 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dw9wt\" (UniqueName: \"kubernetes.io/projected/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-kube-api-access-dw9wt\") pod \"manila-db-create-2dp4k\" (UID: \"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de\") " pod="openstack/manila-db-create-2dp4k" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.645638 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-operator-scripts\") pod \"manila-db-create-2dp4k\" (UID: \"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de\") " pod="openstack/manila-db-create-2dp4k" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.658303 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.660092 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.662935 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.687290 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.689809 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dw9wt\" (UniqueName: \"kubernetes.io/projected/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-kube-api-access-dw9wt\") pod \"manila-db-create-2dp4k\" (UID: \"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de\") " pod="openstack/manila-db-create-2dp4k" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.738441 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.759880 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6c7b8ff54f-b242w"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.763621 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.771491 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.771643 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.771728 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-z7k4n" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.771880 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779496 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779556 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmtdn\" (UniqueName: \"kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-kube-api-access-bmtdn\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779584 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779604 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-twq6r\" (UniqueName: \"kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-kube-api-access-twq6r\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779624 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779642 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779662 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779693 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779770 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-ceph\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779795 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779831 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-config-data\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779872 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-logs\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779904 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779927 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779943 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779960 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-ceph\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779972 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-logs\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.779989 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-scripts\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.798945 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c7b8ff54f-b242w"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.822796 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-6db2-account-create-update-r8jpd"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.824638 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-6db2-account-create-update-r8jpd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.829259 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.833941 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.847068 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-6db2-account-create-update-r8jpd"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.865792 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2dp4k" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.881638 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.881728 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3368ce70-bad5-4535-a666-ca5b751967d0-operator-scripts\") pod \"manila-6db2-account-create-update-r8jpd\" (UID: \"3368ce70-bad5-4535-a666-ca5b751967d0\") " pod="openstack/manila-6db2-account-create-update-r8jpd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.881766 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00092e4b-4bb9-47bc-83e2-4fde7fe32313-logs\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.881792 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvjp7\" (UniqueName: \"kubernetes.io/projected/3368ce70-bad5-4535-a666-ca5b751967d0-kube-api-access-kvjp7\") pod \"manila-6db2-account-create-update-r8jpd\" (UID: \"3368ce70-bad5-4535-a666-ca5b751967d0\") " pod="openstack/manila-6db2-account-create-update-r8jpd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.881835 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-ceph\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.881860 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.881899 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-scripts\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.881920 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-config-data\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.881948 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-logs\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.881967 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00092e4b-4bb9-47bc-83e2-4fde7fe32313-horizon-secret-key\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882012 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882591 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882621 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882639 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-ceph\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882653 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-logs\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882672 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-scripts\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882691 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svbrp\" (UniqueName: \"kubernetes.io/projected/00092e4b-4bb9-47bc-83e2-4fde7fe32313-kube-api-access-svbrp\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882714 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882755 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmtdn\" (UniqueName: \"kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-kube-api-access-bmtdn\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882772 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-config-data\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882803 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882822 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-twq6r\" (UniqueName: \"kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-kube-api-access-twq6r\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882838 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882853 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.882870 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.884815 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.885323 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.885375 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.886252 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.886579 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-logs\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.888278 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.888530 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-logs\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.892646 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: E1203 18:30:41.894464 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[ceph combined-ca-bundle config-data glance internal-tls-certs kube-api-access-bmtdn logs scripts], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/glance-default-internal-api-0" podUID="58908fed-6ba5-4405-a653-319cd06205a5" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.894806 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-scripts\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.896425 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-scripts\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.897946 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.898389 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-config-data\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.899393 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-7df97bdcbf-fgcfd"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.900720 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.901689 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.910066 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7df97bdcbf-fgcfd"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.911525 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.911532 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-ceph\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.919309 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-ceph\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.919901 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-config-data\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.920675 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-twq6r\" (UniqueName: \"kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-kube-api-access-twq6r\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.921387 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmtdn\" (UniqueName: \"kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-kube-api-access-bmtdn\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.953008 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.953189 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.969191 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.970698 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.986830 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-scripts\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.986926 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00092e4b-4bb9-47bc-83e2-4fde7fe32313-horizon-secret-key\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.986995 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nf8d9\" (UniqueName: \"kubernetes.io/projected/f57a600a-addf-458b-b146-8b294fb7fbc1-kube-api-access-nf8d9\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.987060 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svbrp\" (UniqueName: \"kubernetes.io/projected/00092e4b-4bb9-47bc-83e2-4fde7fe32313-kube-api-access-svbrp\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.987106 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-config-data\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.987130 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f57a600a-addf-458b-b146-8b294fb7fbc1-logs\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.987154 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-config-data\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.987190 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-scripts\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.987279 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3368ce70-bad5-4535-a666-ca5b751967d0-operator-scripts\") pod \"manila-6db2-account-create-update-r8jpd\" (UID: \"3368ce70-bad5-4535-a666-ca5b751967d0\") " pod="openstack/manila-6db2-account-create-update-r8jpd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.987301 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f57a600a-addf-458b-b146-8b294fb7fbc1-horizon-secret-key\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.987342 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00092e4b-4bb9-47bc-83e2-4fde7fe32313-logs\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.987371 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvjp7\" (UniqueName: \"kubernetes.io/projected/3368ce70-bad5-4535-a666-ca5b751967d0-kube-api-access-kvjp7\") pod \"manila-6db2-account-create-update-r8jpd\" (UID: \"3368ce70-bad5-4535-a666-ca5b751967d0\") " pod="openstack/manila-6db2-account-create-update-r8jpd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.988303 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-scripts\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.992690 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00092e4b-4bb9-47bc-83e2-4fde7fe32313-horizon-secret-key\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.993724 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3368ce70-bad5-4535-a666-ca5b751967d0-operator-scripts\") pod \"manila-6db2-account-create-update-r8jpd\" (UID: \"3368ce70-bad5-4535-a666-ca5b751967d0\") " pod="openstack/manila-6db2-account-create-update-r8jpd" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.994188 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00092e4b-4bb9-47bc-83e2-4fde7fe32313-logs\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.995586 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-config-data\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:41 crc kubenswrapper[4787]: I1203 18:30:41.997340 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.020959 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvjp7\" (UniqueName: \"kubernetes.io/projected/3368ce70-bad5-4535-a666-ca5b751967d0-kube-api-access-kvjp7\") pod \"manila-6db2-account-create-update-r8jpd\" (UID: \"3368ce70-bad5-4535-a666-ca5b751967d0\") " pod="openstack/manila-6db2-account-create-update-r8jpd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.021714 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svbrp\" (UniqueName: \"kubernetes.io/projected/00092e4b-4bb9-47bc-83e2-4fde7fe32313-kube-api-access-svbrp\") pod \"horizon-6c7b8ff54f-b242w\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.091115 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nf8d9\" (UniqueName: \"kubernetes.io/projected/f57a600a-addf-458b-b146-8b294fb7fbc1-kube-api-access-nf8d9\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.091195 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-config-data\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.091216 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f57a600a-addf-458b-b146-8b294fb7fbc1-logs\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.091247 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-scripts\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.091308 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f57a600a-addf-458b-b146-8b294fb7fbc1-horizon-secret-key\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.092151 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f57a600a-addf-458b-b146-8b294fb7fbc1-logs\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.093672 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-config-data\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.094093 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-scripts\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.096489 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.104994 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f57a600a-addf-458b-b146-8b294fb7fbc1-horizon-secret-key\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.133301 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nf8d9\" (UniqueName: \"kubernetes.io/projected/f57a600a-addf-458b-b146-8b294fb7fbc1-kube-api-access-nf8d9\") pod \"horizon-7df97bdcbf-fgcfd\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.179068 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-6db2-account-create-update-r8jpd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.400467 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.427321 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-2dp4k"] Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.601810 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:30:42 crc kubenswrapper[4787]: W1203 18:30:42.677800 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbb6d641f_0ab2_4ce6_aff0_407aecdab7e9.slice/crio-f190b8a160b065476761ba40c7f95e6d11c6e7936761b72ecec7dc447e19fc37 WatchSource:0}: Error finding container f190b8a160b065476761ba40c7f95e6d11c6e7936761b72ecec7dc447e19fc37: Status 404 returned error can't find the container with id f190b8a160b065476761ba40c7f95e6d11c6e7936761b72ecec7dc447e19fc37 Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.686386 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-6db2-account-create-update-r8jpd"] Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.727572 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6c7b8ff54f-b242w"] Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.732876 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"77e12136-eaca-4c6a-9346-da3325061b57","Type":"ContainerStarted","Data":"d79d2719138cb00b10da6d9806bb10aaadf9e9925223aec751dcfec9a6282347"} Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.744875 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.746166 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9","Type":"ContainerStarted","Data":"f190b8a160b065476761ba40c7f95e6d11c6e7936761b72ecec7dc447e19fc37"} Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.758137 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.759195 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2dp4k" event={"ID":"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de","Type":"ContainerStarted","Data":"b6fa52c9d53743c709e389673cbdea27248f919653f8f4dca352251f567a4a17"} Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.782178 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.822425 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-logs\") pod \"58908fed-6ba5-4405-a653-319cd06205a5\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.823104 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"58908fed-6ba5-4405-a653-319cd06205a5\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.823282 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-config-data\") pod \"58908fed-6ba5-4405-a653-319cd06205a5\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.823339 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-internal-tls-certs\") pod \"58908fed-6ba5-4405-a653-319cd06205a5\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.823091 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-logs" (OuterVolumeSpecName: "logs") pod "58908fed-6ba5-4405-a653-319cd06205a5" (UID: "58908fed-6ba5-4405-a653-319cd06205a5"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.823468 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-combined-ca-bundle\") pod \"58908fed-6ba5-4405-a653-319cd06205a5\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.823807 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-ceph\") pod \"58908fed-6ba5-4405-a653-319cd06205a5\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.823912 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-httpd-run\") pod \"58908fed-6ba5-4405-a653-319cd06205a5\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.824197 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-scripts\") pod \"58908fed-6ba5-4405-a653-319cd06205a5\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.824394 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bmtdn\" (UniqueName: \"kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-kube-api-access-bmtdn\") pod \"58908fed-6ba5-4405-a653-319cd06205a5\" (UID: \"58908fed-6ba5-4405-a653-319cd06205a5\") " Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.826053 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.829545 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "58908fed-6ba5-4405-a653-319cd06205a5" (UID: "58908fed-6ba5-4405-a653-319cd06205a5"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.835592 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "58908fed-6ba5-4405-a653-319cd06205a5" (UID: "58908fed-6ba5-4405-a653-319cd06205a5"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.839057 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "58908fed-6ba5-4405-a653-319cd06205a5" (UID: "58908fed-6ba5-4405-a653-319cd06205a5"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.839272 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-config-data" (OuterVolumeSpecName: "config-data") pod "58908fed-6ba5-4405-a653-319cd06205a5" (UID: "58908fed-6ba5-4405-a653-319cd06205a5"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.839400 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-kube-api-access-bmtdn" (OuterVolumeSpecName: "kube-api-access-bmtdn") pod "58908fed-6ba5-4405-a653-319cd06205a5" (UID: "58908fed-6ba5-4405-a653-319cd06205a5"). InnerVolumeSpecName "kube-api-access-bmtdn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.840522 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "58908fed-6ba5-4405-a653-319cd06205a5" (UID: "58908fed-6ba5-4405-a653-319cd06205a5"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.840934 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-scripts" (OuterVolumeSpecName: "scripts") pod "58908fed-6ba5-4405-a653-319cd06205a5" (UID: "58908fed-6ba5-4405-a653-319cd06205a5"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.841304 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-ceph" (OuterVolumeSpecName: "ceph") pod "58908fed-6ba5-4405-a653-319cd06205a5" (UID: "58908fed-6ba5-4405-a653-319cd06205a5"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.925428 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-7df97bdcbf-fgcfd"] Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.929585 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.929623 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.929632 4787 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/58908fed-6ba5-4405-a653-319cd06205a5-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.929641 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.929650 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bmtdn\" (UniqueName: \"kubernetes.io/projected/58908fed-6ba5-4405-a653-319cd06205a5-kube-api-access-bmtdn\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.929948 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.929975 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.929991 4787 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58908fed-6ba5-4405-a653-319cd06205a5-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:42 crc kubenswrapper[4787]: W1203 18:30:42.942769 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf57a600a_addf_458b_b146_8b294fb7fbc1.slice/crio-3ddcd9aa090b310667987b521f90eb53d3161782cc265192431e141f120128e8 WatchSource:0}: Error finding container 3ddcd9aa090b310667987b521f90eb53d3161782cc265192431e141f120128e8: Status 404 returned error can't find the container with id 3ddcd9aa090b310667987b521f90eb53d3161782cc265192431e141f120128e8 Dec 03 18:30:42 crc kubenswrapper[4787]: I1203 18:30:42.964406 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.031651 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.824574 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"0f2fd27e-97a7-4019-98fd-1ed092285098","Type":"ContainerStarted","Data":"c7a3a26f0d2a66b5c5ef60119f90a5c735cc94c659169f49986cc8ba9274cf60"} Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.829417 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"77e12136-eaca-4c6a-9346-da3325061b57","Type":"ContainerStarted","Data":"973e5ad49c54f403f0c4d7b7142f5667a3dcbc6910b916dadcdec750bcfc4428"} Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.829470 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"77e12136-eaca-4c6a-9346-da3325061b57","Type":"ContainerStarted","Data":"07546d7feb2663d568a70227063a68dc4e476aae6e15b0b4b6d9c32be1169866"} Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.837325 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9","Type":"ContainerStarted","Data":"ab5c334b16077d05fda3bd24addfcf0e17434d980d4b3538c7129e89c11b346a"} Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.840051 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c7b8ff54f-b242w" event={"ID":"00092e4b-4bb9-47bc-83e2-4fde7fe32313","Type":"ContainerStarted","Data":"c3652d7b5ddf5183448c7a5bc204ba7c2df30e23c40062ca02408e212e846863"} Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.844038 4787 generic.go:334] "Generic (PLEG): container finished" podID="3368ce70-bad5-4535-a666-ca5b751967d0" containerID="1b3628d89d8d0e7add80979b0ba64f0cb7fac1217227a14d6b843dcd250ddbc2" exitCode=0 Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.844094 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-6db2-account-create-update-r8jpd" event={"ID":"3368ce70-bad5-4535-a666-ca5b751967d0","Type":"ContainerDied","Data":"1b3628d89d8d0e7add80979b0ba64f0cb7fac1217227a14d6b843dcd250ddbc2"} Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.844120 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-6db2-account-create-update-r8jpd" event={"ID":"3368ce70-bad5-4535-a666-ca5b751967d0","Type":"ContainerStarted","Data":"c45a2dfb5741bb2dead49443490a8ddd8bca478a203b2cc3898ede510d8fc3c4"} Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.862131 4787 generic.go:334] "Generic (PLEG): container finished" podID="ff52f6c6-c90d-43e7-a0cf-3fb81534f2de" containerID="a62c7a1fd363768b8be82b42194d35c1e13c0e6afcbab83729a12e2d0f85cc0c" exitCode=0 Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.862199 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2dp4k" event={"ID":"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de","Type":"ContainerDied","Data":"a62c7a1fd363768b8be82b42194d35c1e13c0e6afcbab83729a12e2d0f85cc0c"} Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.865948 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:43 crc kubenswrapper[4787]: I1203 18:30:43.866387 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7df97bdcbf-fgcfd" event={"ID":"f57a600a-addf-458b-b146-8b294fb7fbc1","Type":"ContainerStarted","Data":"3ddcd9aa090b310667987b521f90eb53d3161782cc265192431e141f120128e8"} Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.094115 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.105090 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.115434 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.148719 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.233614245 podStartE2EDuration="4.14870113s" podCreationTimestamp="2025-12-03 18:30:40 +0000 UTC" firstStartedPulling="2025-12-03 18:30:41.806805586 +0000 UTC m=+4658.624276545" lastFinishedPulling="2025-12-03 18:30:42.721892471 +0000 UTC m=+4659.539363430" observedRunningTime="2025-12-03 18:30:44.047774765 +0000 UTC m=+4660.865245744" watchObservedRunningTime="2025-12-03 18:30:44.14870113 +0000 UTC m=+4660.966172079" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.151716 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.151841 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.155500 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.155645 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.280487 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-ceph\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.281010 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-logs\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.281249 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wdrbv\" (UniqueName: \"kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-kube-api-access-wdrbv\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.281393 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.281512 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.281635 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.281813 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.281945 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.282081 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.385530 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-ceph\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.385623 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-logs\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.385672 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wdrbv\" (UniqueName: \"kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-kube-api-access-wdrbv\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.385714 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.385804 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.385862 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.385966 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.386043 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.386084 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.386860 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.388001 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.388747 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-logs\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.399779 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-scripts\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.401290 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.405170 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wdrbv\" (UniqueName: \"kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-kube-api-access-wdrbv\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.407747 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-config-data\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.418116 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-ceph\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.418881 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.515764 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.695981 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c7b8ff54f-b242w"] Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.730079 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-66fb6d975b-t28ks"] Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.731961 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.733909 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.785269 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66fb6d975b-t28ks"] Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.813178 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.865661 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7df97bdcbf-fgcfd"] Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.876066 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.885052 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"0f2fd27e-97a7-4019-98fd-1ed092285098","Type":"ContainerStarted","Data":"a033c640e697cbcb8d49afbd3e3a9f6045d45d36918dec2a2278243f3594c54b"} Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.885116 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"0f2fd27e-97a7-4019-98fd-1ed092285098","Type":"ContainerStarted","Data":"0fe5a31ca20e8ac99bcdf210f7b8c6b63b3f0b190e22c4b868b1df481ed29558"} Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.893093 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9","Type":"ContainerStarted","Data":"d045e0ecf729536599a08cd358d904b4da6972cb71142d06aaf5ce4f16ba7fd4"} Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.893357 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" containerName="glance-log" containerID="cri-o://ab5c334b16077d05fda3bd24addfcf0e17434d980d4b3538c7129e89c11b346a" gracePeriod=30 Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.894163 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" containerName="glance-httpd" containerID="cri-o://d045e0ecf729536599a08cd358d904b4da6972cb71142d06aaf5ce4f16ba7fd4" gracePeriod=30 Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.896261 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-5d686989bb-9689v"] Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.901790 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-secret-key\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.901882 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-scripts\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.901958 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-config-data\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.901974 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-tls-certs\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.902000 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-combined-ca-bundle\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.902092 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kw5tk\" (UniqueName: \"kubernetes.io/projected/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-kube-api-access-kw5tk\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.902131 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-logs\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.907733 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.941404 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d686989bb-9689v"] Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.944654 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=4.143573423 podStartE2EDuration="4.944634694s" podCreationTimestamp="2025-12-03 18:30:40 +0000 UTC" firstStartedPulling="2025-12-03 18:30:42.80127327 +0000 UTC m=+4659.618744229" lastFinishedPulling="2025-12-03 18:30:43.602334541 +0000 UTC m=+4660.419805500" observedRunningTime="2025-12-03 18:30:44.909972548 +0000 UTC m=+4661.727443507" watchObservedRunningTime="2025-12-03 18:30:44.944634694 +0000 UTC m=+4661.762105653" Dec 03 18:30:44 crc kubenswrapper[4787]: I1203 18:30:44.986159 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.986145083 podStartE2EDuration="4.986145083s" podCreationTimestamp="2025-12-03 18:30:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:30:44.972593301 +0000 UTC m=+4661.790064260" watchObservedRunningTime="2025-12-03 18:30:44.986145083 +0000 UTC m=+4661.803616042" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004238 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99c71646-510d-4f03-9308-4d0a9ed3c854-logs\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004333 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kv58p\" (UniqueName: \"kubernetes.io/projected/99c71646-510d-4f03-9308-4d0a9ed3c854-kube-api-access-kv58p\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004361 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kw5tk\" (UniqueName: \"kubernetes.io/projected/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-kube-api-access-kw5tk\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004406 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99c71646-510d-4f03-9308-4d0a9ed3c854-config-data\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004447 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-logs\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004484 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-secret-key\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004517 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99c71646-510d-4f03-9308-4d0a9ed3c854-scripts\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004534 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c71646-510d-4f03-9308-4d0a9ed3c854-combined-ca-bundle\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004554 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99c71646-510d-4f03-9308-4d0a9ed3c854-horizon-secret-key\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004590 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-scripts\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004651 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c71646-510d-4f03-9308-4d0a9ed3c854-horizon-tls-certs\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004694 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-config-data\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004710 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-tls-certs\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.004742 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-combined-ca-bundle\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.011793 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-logs\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.012103 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-config-data\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.012312 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-scripts\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.106839 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99c71646-510d-4f03-9308-4d0a9ed3c854-logs\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.107431 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99c71646-510d-4f03-9308-4d0a9ed3c854-logs\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.110257 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kv58p\" (UniqueName: \"kubernetes.io/projected/99c71646-510d-4f03-9308-4d0a9ed3c854-kube-api-access-kv58p\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.110369 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99c71646-510d-4f03-9308-4d0a9ed3c854-config-data\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.110736 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99c71646-510d-4f03-9308-4d0a9ed3c854-scripts\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.114561 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c71646-510d-4f03-9308-4d0a9ed3c854-combined-ca-bundle\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.123610 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99c71646-510d-4f03-9308-4d0a9ed3c854-horizon-secret-key\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.112359 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/99c71646-510d-4f03-9308-4d0a9ed3c854-config-data\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.111309 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/99c71646-510d-4f03-9308-4d0a9ed3c854-scripts\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.124069 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c71646-510d-4f03-9308-4d0a9ed3c854-horizon-tls-certs\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.418539 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-secret-key\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.418686 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-tls-certs\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.419179 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-combined-ca-bundle\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.420219 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kw5tk\" (UniqueName: \"kubernetes.io/projected/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-kube-api-access-kw5tk\") pod \"horizon-66fb6d975b-t28ks\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.425315 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/99c71646-510d-4f03-9308-4d0a9ed3c854-horizon-secret-key\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.425525 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/99c71646-510d-4f03-9308-4d0a9ed3c854-combined-ca-bundle\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.425588 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/99c71646-510d-4f03-9308-4d0a9ed3c854-horizon-tls-certs\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.429173 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kv58p\" (UniqueName: \"kubernetes.io/projected/99c71646-510d-4f03-9308-4d0a9ed3c854-kube-api-access-kv58p\") pod \"horizon-5d686989bb-9689v\" (UID: \"99c71646-510d-4f03-9308-4d0a9ed3c854\") " pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.656111 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.669530 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-6db2-account-create-update-r8jpd" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.689081 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.787217 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="58908fed-6ba5-4405-a653-319cd06205a5" path="/var/lib/kubelet/pods/58908fed-6ba5-4405-a653-319cd06205a5/volumes" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.795435 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2dp4k" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.851338 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3368ce70-bad5-4535-a666-ca5b751967d0-operator-scripts\") pod \"3368ce70-bad5-4535-a666-ca5b751967d0\" (UID: \"3368ce70-bad5-4535-a666-ca5b751967d0\") " Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.851849 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3368ce70-bad5-4535-a666-ca5b751967d0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3368ce70-bad5-4535-a666-ca5b751967d0" (UID: "3368ce70-bad5-4535-a666-ca5b751967d0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.851903 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvjp7\" (UniqueName: \"kubernetes.io/projected/3368ce70-bad5-4535-a666-ca5b751967d0-kube-api-access-kvjp7\") pod \"3368ce70-bad5-4535-a666-ca5b751967d0\" (UID: \"3368ce70-bad5-4535-a666-ca5b751967d0\") " Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.853951 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3368ce70-bad5-4535-a666-ca5b751967d0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.864599 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3368ce70-bad5-4535-a666-ca5b751967d0-kube-api-access-kvjp7" (OuterVolumeSpecName: "kube-api-access-kvjp7") pod "3368ce70-bad5-4535-a666-ca5b751967d0" (UID: "3368ce70-bad5-4535-a666-ca5b751967d0"). InnerVolumeSpecName "kube-api-access-kvjp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.906699 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-6db2-account-create-update-r8jpd" event={"ID":"3368ce70-bad5-4535-a666-ca5b751967d0","Type":"ContainerDied","Data":"c45a2dfb5741bb2dead49443490a8ddd8bca478a203b2cc3898ede510d8fc3c4"} Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.906736 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c45a2dfb5741bb2dead49443490a8ddd8bca478a203b2cc3898ede510d8fc3c4" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.906789 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-6db2-account-create-update-r8jpd" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.914216 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.918909 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-2dp4k" event={"ID":"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de","Type":"ContainerDied","Data":"b6fa52c9d53743c709e389673cbdea27248f919653f8f4dca352251f567a4a17"} Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.918946 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b6fa52c9d53743c709e389673cbdea27248f919653f8f4dca352251f567a4a17" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.919059 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-2dp4k" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.922804 4787 generic.go:334] "Generic (PLEG): container finished" podID="bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" containerID="d045e0ecf729536599a08cd358d904b4da6972cb71142d06aaf5ce4f16ba7fd4" exitCode=0 Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.922823 4787 generic.go:334] "Generic (PLEG): container finished" podID="bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" containerID="ab5c334b16077d05fda3bd24addfcf0e17434d980d4b3538c7129e89c11b346a" exitCode=143 Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.923579 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9","Type":"ContainerDied","Data":"d045e0ecf729536599a08cd358d904b4da6972cb71142d06aaf5ce4f16ba7fd4"} Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.923605 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9","Type":"ContainerDied","Data":"ab5c334b16077d05fda3bd24addfcf0e17434d980d4b3538c7129e89c11b346a"} Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.956071 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-operator-scripts\") pod \"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de\" (UID: \"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de\") " Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.956441 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dw9wt\" (UniqueName: \"kubernetes.io/projected/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-kube-api-access-dw9wt\") pod \"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de\" (UID: \"ff52f6c6-c90d-43e7-a0cf-3fb81534f2de\") " Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.956944 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvjp7\" (UniqueName: \"kubernetes.io/projected/3368ce70-bad5-4535-a666-ca5b751967d0-kube-api-access-kvjp7\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.957857 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ff52f6c6-c90d-43e7-a0cf-3fb81534f2de" (UID: "ff52f6c6-c90d-43e7-a0cf-3fb81534f2de"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:30:45 crc kubenswrapper[4787]: I1203 18:30:45.960200 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-kube-api-access-dw9wt" (OuterVolumeSpecName: "kube-api-access-dw9wt") pod "ff52f6c6-c90d-43e7-a0cf-3fb81534f2de" (UID: "ff52f6c6-c90d-43e7-a0cf-3fb81534f2de"). InnerVolumeSpecName "kube-api-access-dw9wt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.017127 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.029860 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.059799 4787 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.059826 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dw9wt\" (UniqueName: \"kubernetes.io/projected/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de-kube-api-access-dw9wt\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.210427 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-5d686989bb-9689v"] Dec 03 18:30:46 crc kubenswrapper[4787]: W1203 18:30:46.217957 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99c71646_510d_4f03_9308_4d0a9ed3c854.slice/crio-9ccfcba11f916562fec2772c8b2ba1d8e20b18ad98b3eda84d5208cf2802f1bb WatchSource:0}: Error finding container 9ccfcba11f916562fec2772c8b2ba1d8e20b18ad98b3eda84d5208cf2802f1bb: Status 404 returned error can't find the container with id 9ccfcba11f916562fec2772c8b2ba1d8e20b18ad98b3eda84d5208cf2802f1bb Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.344941 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-66fb6d975b-t28ks"] Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.471753 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.576886 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-httpd-run\") pod \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.576921 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-ceph\") pod \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.577079 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-config-data\") pod \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.577106 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.577201 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-logs\") pod \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.577229 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-twq6r\" (UniqueName: \"kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-kube-api-access-twq6r\") pod \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.577331 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-public-tls-certs\") pod \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.577354 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-scripts\") pod \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.577391 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-combined-ca-bundle\") pod \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\" (UID: \"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9\") " Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.578386 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-logs" (OuterVolumeSpecName: "logs") pod "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" (UID: "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.579085 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.580183 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" (UID: "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.680739 4787 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.936884 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d686989bb-9689v" event={"ID":"99c71646-510d-4f03-9308-4d0a9ed3c854","Type":"ContainerStarted","Data":"9ccfcba11f916562fec2772c8b2ba1d8e20b18ad98b3eda84d5208cf2802f1bb"} Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.944522 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"bb6d641f-0ab2-4ce6-aff0-407aecdab7e9","Type":"ContainerDied","Data":"f190b8a160b065476761ba40c7f95e6d11c6e7936761b72ecec7dc447e19fc37"} Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.944576 4787 scope.go:117] "RemoveContainer" containerID="d045e0ecf729536599a08cd358d904b4da6972cb71142d06aaf5ce4f16ba7fd4" Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.944696 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.949808 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fb6d975b-t28ks" event={"ID":"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3","Type":"ContainerStarted","Data":"a30fe7537ceae9aafa3e8b2865cffd07d223f0dcc63e2db857e324409195fa9b"} Dec 03 18:30:46 crc kubenswrapper[4787]: I1203 18:30:46.952395 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b70d895-0004-4c49-b230-1b4f10cbfb95","Type":"ContainerStarted","Data":"79db7f441453faa12fed4899ee016de0f5895fac5bc15bad029b39bf4858f638"} Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.113940 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" (UID: "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.115623 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-ceph" (OuterVolumeSpecName: "ceph") pod "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" (UID: "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.116176 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-scripts" (OuterVolumeSpecName: "scripts") pod "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" (UID: "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.116401 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-kube-api-access-twq6r" (OuterVolumeSpecName: "kube-api-access-twq6r") pod "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" (UID: "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9"). InnerVolumeSpecName "kube-api-access-twq6r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.197509 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.197545 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-twq6r\" (UniqueName: \"kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-kube-api-access-twq6r\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.197560 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.197571 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.240443 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" (UID: "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.300906 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.341882 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.375235 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-config-data" (OuterVolumeSpecName: "config-data") pod "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" (UID: "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.380205 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" (UID: "bb6d641f-0ab2-4ce6-aff0-407aecdab7e9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.409645 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.409683 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.409716 4787 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.512656 4787 scope.go:117] "RemoveContainer" containerID="ab5c334b16077d05fda3bd24addfcf0e17434d980d4b3538c7129e89c11b346a" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.692389 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.713632 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.728224 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:30:47 crc kubenswrapper[4787]: E1203 18:30:47.728699 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" containerName="glance-log" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.728724 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" containerName="glance-log" Dec 03 18:30:47 crc kubenswrapper[4787]: E1203 18:30:47.728757 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" containerName="glance-httpd" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.728767 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" containerName="glance-httpd" Dec 03 18:30:47 crc kubenswrapper[4787]: E1203 18:30:47.728795 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ff52f6c6-c90d-43e7-a0cf-3fb81534f2de" containerName="mariadb-database-create" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.728801 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ff52f6c6-c90d-43e7-a0cf-3fb81534f2de" containerName="mariadb-database-create" Dec 03 18:30:47 crc kubenswrapper[4787]: E1203 18:30:47.728823 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3368ce70-bad5-4535-a666-ca5b751967d0" containerName="mariadb-account-create-update" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.728829 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3368ce70-bad5-4535-a666-ca5b751967d0" containerName="mariadb-account-create-update" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.729039 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" containerName="glance-httpd" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.729058 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ff52f6c6-c90d-43e7-a0cf-3fb81534f2de" containerName="mariadb-database-create" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.729069 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" containerName="glance-log" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.729080 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="3368ce70-bad5-4535-a666-ca5b751967d0" containerName="mariadb-account-create-update" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.730271 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.733638 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.733815 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.753472 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.768254 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:30:47 crc kubenswrapper[4787]: E1203 18:30:47.768561 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.791996 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bb6d641f-0ab2-4ce6-aff0-407aecdab7e9" path="/var/lib/kubelet/pods/bb6d641f-0ab2-4ce6-aff0-407aecdab7e9/volumes" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.920912 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.921145 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.921179 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62231118-050c-48ab-9013-0c07ad92cb6b-logs\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.921214 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwm85\" (UniqueName: \"kubernetes.io/projected/62231118-050c-48ab-9013-0c07ad92cb6b-kube-api-access-pwm85\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.921268 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.921285 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/62231118-050c-48ab-9013-0c07ad92cb6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.921336 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.921393 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.921450 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/62231118-050c-48ab-9013-0c07ad92cb6b-ceph\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:47 crc kubenswrapper[4787]: I1203 18:30:47.967985 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b70d895-0004-4c49-b230-1b4f10cbfb95","Type":"ContainerStarted","Data":"4176d0827b508351b25e9ec448443e977841d0439df2256b90274ecf67b6d004"} Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.035340 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.035441 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.035511 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/62231118-050c-48ab-9013-0c07ad92cb6b-ceph\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.035586 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.035692 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.035723 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62231118-050c-48ab-9013-0c07ad92cb6b-logs\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.035757 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwm85\" (UniqueName: \"kubernetes.io/projected/62231118-050c-48ab-9013-0c07ad92cb6b-kube-api-access-pwm85\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.035817 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.035836 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/62231118-050c-48ab-9013-0c07ad92cb6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.036403 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/62231118-050c-48ab-9013-0c07ad92cb6b-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.036988 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/62231118-050c-48ab-9013-0c07ad92cb6b-logs\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.038081 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.043220 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.043540 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-scripts\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.047708 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/62231118-050c-48ab-9013-0c07ad92cb6b-ceph\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.055520 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.055571 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwm85\" (UniqueName: \"kubernetes.io/projected/62231118-050c-48ab-9013-0c07ad92cb6b-kube-api-access-pwm85\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.066882 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/62231118-050c-48ab-9013-0c07ad92cb6b-config-data\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.085679 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-0\" (UID: \"62231118-050c-48ab-9013-0c07ad92cb6b\") " pod="openstack/glance-default-external-api-0" Dec 03 18:30:48 crc kubenswrapper[4787]: I1203 18:30:48.350188 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 03 18:30:50 crc kubenswrapper[4787]: E1203 18:30:50.015525 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69ab81e9_ff7c_40f4_9f76_6f9e06e4caee.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice/crio-98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6\": RecentStats: unable to find data in memory cache]" Dec 03 18:30:51 crc kubenswrapper[4787]: I1203 18:30:51.129808 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 03 18:30:51 crc kubenswrapper[4787]: I1203 18:30:51.264777 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 03 18:30:51 crc kubenswrapper[4787]: I1203 18:30:51.992224 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-bfb6h"] Dec 03 18:30:51 crc kubenswrapper[4787]: I1203 18:30:51.994260 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:51 crc kubenswrapper[4787]: I1203 18:30:51.996850 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-wlvhv" Dec 03 18:30:51 crc kubenswrapper[4787]: I1203 18:30:51.999072 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.004296 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-bfb6h"] Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.039180 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-config-data\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.039281 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-combined-ca-bundle\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.039471 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9tzl\" (UniqueName: \"kubernetes.io/projected/b8e3c8d9-6215-4247-a39a-27a99e53a33f-kube-api-access-g9tzl\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.039640 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-job-config-data\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.141885 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9tzl\" (UniqueName: \"kubernetes.io/projected/b8e3c8d9-6215-4247-a39a-27a99e53a33f-kube-api-access-g9tzl\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.141981 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-job-config-data\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.142081 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-config-data\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.142113 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-combined-ca-bundle\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.149587 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-combined-ca-bundle\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.151612 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-config-data\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.155969 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-job-config-data\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.160986 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9tzl\" (UniqueName: \"kubernetes.io/projected/b8e3c8d9-6215-4247-a39a-27a99e53a33f-kube-api-access-g9tzl\") pod \"manila-db-sync-bfb6h\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:52 crc kubenswrapper[4787]: I1203 18:30:52.327031 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-bfb6h" Dec 03 18:30:54 crc kubenswrapper[4787]: I1203 18:30:54.057975 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 03 18:30:54 crc kubenswrapper[4787]: I1203 18:30:54.060107 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c7b8ff54f-b242w" event={"ID":"00092e4b-4bb9-47bc-83e2-4fde7fe32313","Type":"ContainerStarted","Data":"50e484e7525c15bfd0b2aeede9f32e007d3d99a162172b8e29eb7dcea9553f17"} Dec 03 18:30:54 crc kubenswrapper[4787]: I1203 18:30:54.064034 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d686989bb-9689v" event={"ID":"99c71646-510d-4f03-9308-4d0a9ed3c854","Type":"ContainerStarted","Data":"be3a41e239a0c0b087bc56329c8d61effd711cc2ca6578fe03ee657c401794e3"} Dec 03 18:30:54 crc kubenswrapper[4787]: I1203 18:30:54.068918 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7df97bdcbf-fgcfd" event={"ID":"f57a600a-addf-458b-b146-8b294fb7fbc1","Type":"ContainerStarted","Data":"2f3b8125d69649225b1ee59bbdf1301d30207516a0d4fecc019b817167db5047"} Dec 03 18:30:54 crc kubenswrapper[4787]: I1203 18:30:54.071761 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fb6d975b-t28ks" event={"ID":"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3","Type":"ContainerStarted","Data":"422f921f2714b4435cd12b9578e5ee5f5f35831b82e8bae5717d3c307be5667b"} Dec 03 18:30:54 crc kubenswrapper[4787]: I1203 18:30:54.177653 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-bfb6h"] Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.086418 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b70d895-0004-4c49-b230-1b4f10cbfb95","Type":"ContainerStarted","Data":"20876c5e340b2f1740162f0fd9c184da60ed641b63e778e862a27a77156018f3"} Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.086488 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6b70d895-0004-4c49-b230-1b4f10cbfb95" containerName="glance-log" containerID="cri-o://4176d0827b508351b25e9ec448443e977841d0439df2256b90274ecf67b6d004" gracePeriod=30 Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.086548 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="6b70d895-0004-4c49-b230-1b4f10cbfb95" containerName="glance-httpd" containerID="cri-o://20876c5e340b2f1740162f0fd9c184da60ed641b63e778e862a27a77156018f3" gracePeriod=30 Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.099753 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-5d686989bb-9689v" event={"ID":"99c71646-510d-4f03-9308-4d0a9ed3c854","Type":"ContainerStarted","Data":"417f8d8c223fbb233f116e248d823428e0dc83ce9c4af3f479c2d382b5cab9ca"} Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.102598 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7df97bdcbf-fgcfd" event={"ID":"f57a600a-addf-458b-b146-8b294fb7fbc1","Type":"ContainerStarted","Data":"f27aba4fd6dbbc751ecab3a1bd1df6ec5e5e441eb49b4abe6e28358a091ecb41"} Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.102729 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7df97bdcbf-fgcfd" podUID="f57a600a-addf-458b-b146-8b294fb7fbc1" containerName="horizon-log" containerID="cri-o://2f3b8125d69649225b1ee59bbdf1301d30207516a0d4fecc019b817167db5047" gracePeriod=30 Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.102992 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-7df97bdcbf-fgcfd" podUID="f57a600a-addf-458b-b146-8b294fb7fbc1" containerName="horizon" containerID="cri-o://f27aba4fd6dbbc751ecab3a1bd1df6ec5e5e441eb49b4abe6e28358a091ecb41" gracePeriod=30 Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.120715 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fb6d975b-t28ks" event={"ID":"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3","Type":"ContainerStarted","Data":"4de5b41a3df3c3e88ba31d7984a50c7e78af565c2be91b7f608830527d0c964a"} Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.125308 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=11.125258132 podStartE2EDuration="11.125258132s" podCreationTimestamp="2025-12-03 18:30:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:30:55.119488077 +0000 UTC m=+4671.936959036" watchObservedRunningTime="2025-12-03 18:30:55.125258132 +0000 UTC m=+4671.942729091" Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.127318 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"62231118-050c-48ab-9013-0c07ad92cb6b","Type":"ContainerStarted","Data":"90af0374849324c94fb2ea6967c8f7059085fc92a1633565b284cc8a9b1365c1"} Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.127432 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"62231118-050c-48ab-9013-0c07ad92cb6b","Type":"ContainerStarted","Data":"55655f59287a0490568c1a8aafe4d8c5f094d64b3d49bb89d9eab5988811b0ab"} Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.129436 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6c7b8ff54f-b242w" podUID="00092e4b-4bb9-47bc-83e2-4fde7fe32313" containerName="horizon-log" containerID="cri-o://50e484e7525c15bfd0b2aeede9f32e007d3d99a162172b8e29eb7dcea9553f17" gracePeriod=30 Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.129697 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c7b8ff54f-b242w" event={"ID":"00092e4b-4bb9-47bc-83e2-4fde7fe32313","Type":"ContainerStarted","Data":"d4095d3a681b531029bf766106b6018170fabda894ecb834035206ad413f2afc"} Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.129754 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-6c7b8ff54f-b242w" podUID="00092e4b-4bb9-47bc-83e2-4fde7fe32313" containerName="horizon" containerID="cri-o://d4095d3a681b531029bf766106b6018170fabda894ecb834035206ad413f2afc" gracePeriod=30 Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.134814 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-bfb6h" event={"ID":"b8e3c8d9-6215-4247-a39a-27a99e53a33f","Type":"ContainerStarted","Data":"5b867b1f1e38c74f23a910e5e4dc6f79c13fb0d2380c808434257c070230e5af"} Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.145936 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-5d686989bb-9689v" podStartSLOduration=3.86933625 podStartE2EDuration="11.145917143s" podCreationTimestamp="2025-12-03 18:30:44 +0000 UTC" firstStartedPulling="2025-12-03 18:30:46.234498226 +0000 UTC m=+4663.051969185" lastFinishedPulling="2025-12-03 18:30:53.511079119 +0000 UTC m=+4670.328550078" observedRunningTime="2025-12-03 18:30:55.135061563 +0000 UTC m=+4671.952532532" watchObservedRunningTime="2025-12-03 18:30:55.145917143 +0000 UTC m=+4671.963388102" Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.171346 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-7df97bdcbf-fgcfd" podStartSLOduration=3.601838641 podStartE2EDuration="14.171331692s" podCreationTimestamp="2025-12-03 18:30:41 +0000 UTC" firstStartedPulling="2025-12-03 18:30:42.953822044 +0000 UTC m=+4659.771293003" lastFinishedPulling="2025-12-03 18:30:53.523315095 +0000 UTC m=+4670.340786054" observedRunningTime="2025-12-03 18:30:55.165457485 +0000 UTC m=+4671.982928444" watchObservedRunningTime="2025-12-03 18:30:55.171331692 +0000 UTC m=+4671.988802651" Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.194458 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-66fb6d975b-t28ks" podStartSLOduration=4.054580737 podStartE2EDuration="11.194442339s" podCreationTimestamp="2025-12-03 18:30:44 +0000 UTC" firstStartedPulling="2025-12-03 18:30:46.383452533 +0000 UTC m=+4663.200923492" lastFinishedPulling="2025-12-03 18:30:53.523314135 +0000 UTC m=+4670.340785094" observedRunningTime="2025-12-03 18:30:55.183917208 +0000 UTC m=+4672.001388167" watchObservedRunningTime="2025-12-03 18:30:55.194442339 +0000 UTC m=+4672.011913298" Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.210230 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6c7b8ff54f-b242w" podStartSLOduration=3.358461122 podStartE2EDuration="14.21020814s" podCreationTimestamp="2025-12-03 18:30:41 +0000 UTC" firstStartedPulling="2025-12-03 18:30:42.735010781 +0000 UTC m=+4659.552481740" lastFinishedPulling="2025-12-03 18:30:53.586757799 +0000 UTC m=+4670.404228758" observedRunningTime="2025-12-03 18:30:55.20311002 +0000 UTC m=+4672.020580989" watchObservedRunningTime="2025-12-03 18:30:55.21020814 +0000 UTC m=+4672.027679099" Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.657094 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.657148 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.690951 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:55 crc kubenswrapper[4787]: I1203 18:30:55.691031 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.147113 4787 generic.go:334] "Generic (PLEG): container finished" podID="6b70d895-0004-4c49-b230-1b4f10cbfb95" containerID="20876c5e340b2f1740162f0fd9c184da60ed641b63e778e862a27a77156018f3" exitCode=0 Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.147399 4787 generic.go:334] "Generic (PLEG): container finished" podID="6b70d895-0004-4c49-b230-1b4f10cbfb95" containerID="4176d0827b508351b25e9ec448443e977841d0439df2256b90274ecf67b6d004" exitCode=143 Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.147180 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b70d895-0004-4c49-b230-1b4f10cbfb95","Type":"ContainerDied","Data":"20876c5e340b2f1740162f0fd9c184da60ed641b63e778e862a27a77156018f3"} Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.147494 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b70d895-0004-4c49-b230-1b4f10cbfb95","Type":"ContainerDied","Data":"4176d0827b508351b25e9ec448443e977841d0439df2256b90274ecf67b6d004"} Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.435217 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.567837 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wdrbv\" (UniqueName: \"kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-kube-api-access-wdrbv\") pod \"6b70d895-0004-4c49-b230-1b4f10cbfb95\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.568300 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-combined-ca-bundle\") pod \"6b70d895-0004-4c49-b230-1b4f10cbfb95\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.568338 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-httpd-run\") pod \"6b70d895-0004-4c49-b230-1b4f10cbfb95\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.568395 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-internal-tls-certs\") pod \"6b70d895-0004-4c49-b230-1b4f10cbfb95\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.568434 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-config-data\") pod \"6b70d895-0004-4c49-b230-1b4f10cbfb95\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.568460 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"6b70d895-0004-4c49-b230-1b4f10cbfb95\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.568509 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-ceph\") pod \"6b70d895-0004-4c49-b230-1b4f10cbfb95\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.568570 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-logs\") pod \"6b70d895-0004-4c49-b230-1b4f10cbfb95\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.568590 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-scripts\") pod \"6b70d895-0004-4c49-b230-1b4f10cbfb95\" (UID: \"6b70d895-0004-4c49-b230-1b4f10cbfb95\") " Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.570055 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-logs" (OuterVolumeSpecName: "logs") pod "6b70d895-0004-4c49-b230-1b4f10cbfb95" (UID: "6b70d895-0004-4c49-b230-1b4f10cbfb95"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.570287 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6b70d895-0004-4c49-b230-1b4f10cbfb95" (UID: "6b70d895-0004-4c49-b230-1b4f10cbfb95"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.609711 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage05-crc" (OuterVolumeSpecName: "glance") pod "6b70d895-0004-4c49-b230-1b4f10cbfb95" (UID: "6b70d895-0004-4c49-b230-1b4f10cbfb95"). InnerVolumeSpecName "local-storage05-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.610630 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-scripts" (OuterVolumeSpecName: "scripts") pod "6b70d895-0004-4c49-b230-1b4f10cbfb95" (UID: "6b70d895-0004-4c49-b230-1b4f10cbfb95"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.611084 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-ceph" (OuterVolumeSpecName: "ceph") pod "6b70d895-0004-4c49-b230-1b4f10cbfb95" (UID: "6b70d895-0004-4c49-b230-1b4f10cbfb95"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.612199 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-kube-api-access-wdrbv" (OuterVolumeSpecName: "kube-api-access-wdrbv") pod "6b70d895-0004-4c49-b230-1b4f10cbfb95" (UID: "6b70d895-0004-4c49-b230-1b4f10cbfb95"). InnerVolumeSpecName "kube-api-access-wdrbv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.616975 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6b70d895-0004-4c49-b230-1b4f10cbfb95" (UID: "6b70d895-0004-4c49-b230-1b4f10cbfb95"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.643224 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "6b70d895-0004-4c49-b230-1b4f10cbfb95" (UID: "6b70d895-0004-4c49-b230-1b4f10cbfb95"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.647333 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-config-data" (OuterVolumeSpecName: "config-data") pod "6b70d895-0004-4c49-b230-1b4f10cbfb95" (UID: "6b70d895-0004-4c49-b230-1b4f10cbfb95"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.671457 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.671485 4787 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-httpd-run\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.671494 4787 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.671502 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.671547 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" " Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.671556 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.671564 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b70d895-0004-4c49-b230-1b4f10cbfb95-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.671573 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6b70d895-0004-4c49-b230-1b4f10cbfb95-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.671582 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wdrbv\" (UniqueName: \"kubernetes.io/projected/6b70d895-0004-4c49-b230-1b4f10cbfb95-kube-api-access-wdrbv\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.698375 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage05-crc" (UniqueName: "kubernetes.io/local-volume/local-storage05-crc") on node "crc" Dec 03 18:30:56 crc kubenswrapper[4787]: I1203 18:30:56.774152 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") on node \"crc\" DevicePath \"\"" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.159630 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"62231118-050c-48ab-9013-0c07ad92cb6b","Type":"ContainerStarted","Data":"47a752f649613df00f94435914417e482595613938c885b0e14d444d66c9157f"} Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.189271 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"6b70d895-0004-4c49-b230-1b4f10cbfb95","Type":"ContainerDied","Data":"79db7f441453faa12fed4899ee016de0f5895fac5bc15bad029b39bf4858f638"} Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.189348 4787 scope.go:117] "RemoveContainer" containerID="20876c5e340b2f1740162f0fd9c184da60ed641b63e778e862a27a77156018f3" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.189555 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.217361 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=10.217332465 podStartE2EDuration="10.217332465s" podCreationTimestamp="2025-12-03 18:30:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:30:57.18946645 +0000 UTC m=+4674.006937409" watchObservedRunningTime="2025-12-03 18:30:57.217332465 +0000 UTC m=+4674.034803424" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.276989 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.305674 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.329025 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:57 crc kubenswrapper[4787]: E1203 18:30:57.329738 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b70d895-0004-4c49-b230-1b4f10cbfb95" containerName="glance-httpd" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.329769 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b70d895-0004-4c49-b230-1b4f10cbfb95" containerName="glance-httpd" Dec 03 18:30:57 crc kubenswrapper[4787]: E1203 18:30:57.329786 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b70d895-0004-4c49-b230-1b4f10cbfb95" containerName="glance-log" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.329792 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b70d895-0004-4c49-b230-1b4f10cbfb95" containerName="glance-log" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.330008 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b70d895-0004-4c49-b230-1b4f10cbfb95" containerName="glance-httpd" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.330133 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b70d895-0004-4c49-b230-1b4f10cbfb95" containerName="glance-log" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.331335 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.339928 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.340635 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.340682 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.502822 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/21ef4218-4f36-423e-abd5-86b398276be8-ceph\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.503470 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21ef4218-4f36-423e-abd5-86b398276be8-logs\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.503563 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.503602 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.504440 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21ef4218-4f36-423e-abd5-86b398276be8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.504519 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.504570 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.504645 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb5vx\" (UniqueName: \"kubernetes.io/projected/21ef4218-4f36-423e-abd5-86b398276be8-kube-api-access-fb5vx\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.507135 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.610572 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/21ef4218-4f36-423e-abd5-86b398276be8-ceph\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.610648 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21ef4218-4f36-423e-abd5-86b398276be8-logs\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.610733 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.610770 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.610788 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21ef4218-4f36-423e-abd5-86b398276be8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.610815 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.610837 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.610864 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb5vx\" (UniqueName: \"kubernetes.io/projected/21ef4218-4f36-423e-abd5-86b398276be8-kube-api-access-fb5vx\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.610885 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.614159 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.616440 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/21ef4218-4f36-423e-abd5-86b398276be8-logs\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.616475 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/21ef4218-4f36-423e-abd5-86b398276be8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:57 crc kubenswrapper[4787]: I1203 18:30:57.786081 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b70d895-0004-4c49-b230-1b4f10cbfb95" path="/var/lib/kubelet/pods/6b70d895-0004-4c49-b230-1b4f10cbfb95/volumes" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.213910 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.213967 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.216759 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/21ef4218-4f36-423e-abd5-86b398276be8-ceph\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.218267 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.220522 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/21ef4218-4f36-423e-abd5-86b398276be8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.223829 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb5vx\" (UniqueName: \"kubernetes.io/projected/21ef4218-4f36-423e-abd5-86b398276be8-kube-api-access-fb5vx\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.254793 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-internal-api-0\" (UID: \"21ef4218-4f36-423e-abd5-86b398276be8\") " pod="openstack/glance-default-internal-api-0" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.349442 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.349671 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.396522 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.449505 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 03 18:30:58 crc kubenswrapper[4787]: I1203 18:30:58.567012 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 03 18:30:59 crc kubenswrapper[4787]: I1203 18:30:59.216819 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 18:30:59 crc kubenswrapper[4787]: I1203 18:30:59.216860 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 03 18:31:00 crc kubenswrapper[4787]: E1203 18:31:00.311471 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod69ab81e9_ff7c_40f4_9f76_6f9e06e4caee.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice/crio-98a791bce44a1a18194fb8336a6b783a4a6cc00d77010f746461fd66c03a7ad6\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod356d1186_2f36_4101_80d0_0d240814bb08.slice\": RecentStats: unable to find data in memory cache]" Dec 03 18:31:00 crc kubenswrapper[4787]: I1203 18:31:00.390413 4787 scope.go:117] "RemoveContainer" containerID="4176d0827b508351b25e9ec448443e977841d0439df2256b90274ecf67b6d004" Dec 03 18:31:00 crc kubenswrapper[4787]: I1203 18:31:00.769152 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:31:00 crc kubenswrapper[4787]: E1203 18:31:00.770336 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:31:01 crc kubenswrapper[4787]: I1203 18:31:01.040913 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 03 18:31:01 crc kubenswrapper[4787]: W1203 18:31:01.047373 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod21ef4218_4f36_423e_abd5_86b398276be8.slice/crio-440ee237b43805d9c734c776611619e33e87676c322cf847d81b1ae5408e3c8c WatchSource:0}: Error finding container 440ee237b43805d9c734c776611619e33e87676c322cf847d81b1ae5408e3c8c: Status 404 returned error can't find the container with id 440ee237b43805d9c734c776611619e33e87676c322cf847d81b1ae5408e3c8c Dec 03 18:31:01 crc kubenswrapper[4787]: I1203 18:31:01.253487 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-bfb6h" event={"ID":"b8e3c8d9-6215-4247-a39a-27a99e53a33f","Type":"ContainerStarted","Data":"d0fa1c521211d166eac35c4a193398805fabcf828e8fae550e75dbde8806cc44"} Dec 03 18:31:01 crc kubenswrapper[4787]: I1203 18:31:01.260233 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21ef4218-4f36-423e-abd5-86b398276be8","Type":"ContainerStarted","Data":"440ee237b43805d9c734c776611619e33e87676c322cf847d81b1ae5408e3c8c"} Dec 03 18:31:01 crc kubenswrapper[4787]: I1203 18:31:01.278004 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-bfb6h" podStartSLOduration=4.018586103 podStartE2EDuration="10.277981184s" podCreationTimestamp="2025-12-03 18:30:51 +0000 UTC" firstStartedPulling="2025-12-03 18:30:54.192925836 +0000 UTC m=+4671.010396795" lastFinishedPulling="2025-12-03 18:31:00.452320917 +0000 UTC m=+4677.269791876" observedRunningTime="2025-12-03 18:31:01.270295249 +0000 UTC m=+4678.087766228" watchObservedRunningTime="2025-12-03 18:31:01.277981184 +0000 UTC m=+4678.095452153" Dec 03 18:31:02 crc kubenswrapper[4787]: I1203 18:31:02.102691 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:31:02 crc kubenswrapper[4787]: I1203 18:31:02.288247 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21ef4218-4f36-423e-abd5-86b398276be8","Type":"ContainerStarted","Data":"f10a8e75123df825332adebee179815164534843d457680c42bf2de4f10cbf43"} Dec 03 18:31:02 crc kubenswrapper[4787]: I1203 18:31:02.288604 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"21ef4218-4f36-423e-abd5-86b398276be8","Type":"ContainerStarted","Data":"edcb2f77bd9d8ce5fbc96cbb1240bfe2a748972a5a5662992b85818c4d495c8e"} Dec 03 18:31:02 crc kubenswrapper[4787]: I1203 18:31:02.328563 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.328545547 podStartE2EDuration="5.328545547s" podCreationTimestamp="2025-12-03 18:30:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:02.309303883 +0000 UTC m=+4679.126774852" watchObservedRunningTime="2025-12-03 18:31:02.328545547 +0000 UTC m=+4679.146016506" Dec 03 18:31:02 crc kubenswrapper[4787]: I1203 18:31:02.396852 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 18:31:02 crc kubenswrapper[4787]: I1203 18:31:02.401646 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:31:02 crc kubenswrapper[4787]: I1203 18:31:02.411788 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 03 18:31:05 crc kubenswrapper[4787]: I1203 18:31:05.660559 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-5d686989bb-9689v" podUID="99c71646-510d-4f03-9308-4d0a9ed3c854" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.72:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.72:8443: connect: connection refused" Dec 03 18:31:05 crc kubenswrapper[4787]: I1203 18:31:05.693559 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-66fb6d975b-t28ks" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.71:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.71:8443: connect: connection refused" Dec 03 18:31:08 crc kubenswrapper[4787]: I1203 18:31:08.567753 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:08 crc kubenswrapper[4787]: I1203 18:31:08.568345 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:09 crc kubenswrapper[4787]: I1203 18:31:09.350121 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:09 crc kubenswrapper[4787]: I1203 18:31:09.368334 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:09 crc kubenswrapper[4787]: I1203 18:31:09.379809 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:09 crc kubenswrapper[4787]: I1203 18:31:09.379871 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:11 crc kubenswrapper[4787]: I1203 18:31:11.343282 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:11 crc kubenswrapper[4787]: I1203 18:31:11.347478 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 03 18:31:11 crc kubenswrapper[4787]: I1203 18:31:11.421879 4787 generic.go:334] "Generic (PLEG): container finished" podID="b8e3c8d9-6215-4247-a39a-27a99e53a33f" containerID="d0fa1c521211d166eac35c4a193398805fabcf828e8fae550e75dbde8806cc44" exitCode=0 Dec 03 18:31:11 crc kubenswrapper[4787]: I1203 18:31:11.421935 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-bfb6h" event={"ID":"b8e3c8d9-6215-4247-a39a-27a99e53a33f","Type":"ContainerDied","Data":"d0fa1c521211d166eac35c4a193398805fabcf828e8fae550e75dbde8806cc44"} Dec 03 18:31:12 crc kubenswrapper[4787]: I1203 18:31:12.765823 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:31:12 crc kubenswrapper[4787]: E1203 18:31:12.766991 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:31:12 crc kubenswrapper[4787]: I1203 18:31:12.948478 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-bfb6h" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.118326 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-config-data\") pod \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.118512 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-job-config-data\") pod \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.118603 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-combined-ca-bundle\") pod \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.118994 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g9tzl\" (UniqueName: \"kubernetes.io/projected/b8e3c8d9-6215-4247-a39a-27a99e53a33f-kube-api-access-g9tzl\") pod \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\" (UID: \"b8e3c8d9-6215-4247-a39a-27a99e53a33f\") " Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.125506 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "b8e3c8d9-6215-4247-a39a-27a99e53a33f" (UID: "b8e3c8d9-6215-4247-a39a-27a99e53a33f"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.125539 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b8e3c8d9-6215-4247-a39a-27a99e53a33f-kube-api-access-g9tzl" (OuterVolumeSpecName: "kube-api-access-g9tzl") pod "b8e3c8d9-6215-4247-a39a-27a99e53a33f" (UID: "b8e3c8d9-6215-4247-a39a-27a99e53a33f"). InnerVolumeSpecName "kube-api-access-g9tzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.129130 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-config-data" (OuterVolumeSpecName: "config-data") pod "b8e3c8d9-6215-4247-a39a-27a99e53a33f" (UID: "b8e3c8d9-6215-4247-a39a-27a99e53a33f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.172294 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b8e3c8d9-6215-4247-a39a-27a99e53a33f" (UID: "b8e3c8d9-6215-4247-a39a-27a99e53a33f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.221917 4787 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.221953 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.221968 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g9tzl\" (UniqueName: \"kubernetes.io/projected/b8e3c8d9-6215-4247-a39a-27a99e53a33f-kube-api-access-g9tzl\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.221984 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b8e3c8d9-6215-4247-a39a-27a99e53a33f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.444162 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-bfb6h" event={"ID":"b8e3c8d9-6215-4247-a39a-27a99e53a33f","Type":"ContainerDied","Data":"5b867b1f1e38c74f23a910e5e4dc6f79c13fb0d2380c808434257c070230e5af"} Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.444221 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5b867b1f1e38c74f23a910e5e4dc6f79c13fb0d2380c808434257c070230e5af" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.444191 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-bfb6h" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.878348 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:31:13 crc kubenswrapper[4787]: E1203 18:31:13.878827 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b8e3c8d9-6215-4247-a39a-27a99e53a33f" containerName="manila-db-sync" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.878840 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b8e3c8d9-6215-4247-a39a-27a99e53a33f" containerName="manila-db-sync" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.879058 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b8e3c8d9-6215-4247-a39a-27a99e53a33f" containerName="manila-db-sync" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.882539 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.890334 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.890347 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-wlvhv" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.891960 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.892183 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.910813 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.913357 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.919114 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.931192 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:31:13 crc kubenswrapper[4787]: I1203 18:31:13.998488 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.035366 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5767ddb7c-wsddg"] Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.037349 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.043101 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.043146 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.043309 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.043365 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wsvh8\" (UniqueName: \"kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-kube-api-access-wsvh8\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.043435 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.043662 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.044125 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.044149 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-scripts\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.044350 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.044555 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hq722\" (UniqueName: \"kubernetes.io/projected/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-kube-api-access-hq722\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.044600 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-scripts\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.044703 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-ceph\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.044786 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.045958 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.074859 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5767ddb7c-wsddg"] Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.111959 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.127007 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.149446 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.153644 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.153716 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.153746 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.153774 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.153825 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wsvh8\" (UniqueName: \"kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-kube-api-access-wsvh8\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.153853 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-ovsdbserver-sb\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.153916 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.153961 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-config\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.153998 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jsdr\" (UniqueName: \"kubernetes.io/projected/060aca68-9071-46af-a627-ed80ad206d74-kube-api-access-7jsdr\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154087 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-dns-swift-storage-0\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154114 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154135 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-ovsdbserver-nb\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154163 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154187 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-scripts\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154321 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154391 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hq722\" (UniqueName: \"kubernetes.io/projected/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-kube-api-access-hq722\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154418 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-scripts\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154463 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-ceph\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154500 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-openstack-edpm-ipam\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154521 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-dns-svc\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154549 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.154698 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.156697 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.158302 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.161954 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.162122 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.175229 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-ceph\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.179854 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.181296 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.181584 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.192181 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wsvh8\" (UniqueName: \"kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-kube-api-access-wsvh8\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.192517 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-scripts\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.193912 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.202220 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-scripts\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.202804 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hq722\" (UniqueName: \"kubernetes.io/projected/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-kube-api-access-hq722\") pod \"manila-scheduler-0\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.208859 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data\") pod \"manila-share-share1-0\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.213572 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.241992 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.256766 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1c54538-1646-4a87-9240-fba80302af69-etc-machine-id\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.256916 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-openstack-edpm-ipam\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.256941 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-dns-svc\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.256962 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-scripts\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.258926 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-openstack-edpm-ipam\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.259012 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.259246 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data-custom\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.259307 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-ovsdbserver-sb\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.259887 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-config\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.260080 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-config\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.260464 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-dns-svc\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.260653 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-ovsdbserver-sb\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.261241 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jsdr\" (UniqueName: \"kubernetes.io/projected/060aca68-9071-46af-a627-ed80ad206d74-kube-api-access-7jsdr\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.261331 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x44pz\" (UniqueName: \"kubernetes.io/projected/f1c54538-1646-4a87-9240-fba80302af69-kube-api-access-x44pz\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.261434 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1c54538-1646-4a87-9240-fba80302af69-logs\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.261465 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-dns-swift-storage-0\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.262171 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-dns-swift-storage-0\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.262386 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-ovsdbserver-nb\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.262486 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.265365 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/060aca68-9071-46af-a627-ed80ad206d74-ovsdbserver-nb\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.293458 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jsdr\" (UniqueName: \"kubernetes.io/projected/060aca68-9071-46af-a627-ed80ad206d74-kube-api-access-7jsdr\") pod \"dnsmasq-dns-5767ddb7c-wsddg\" (UID: \"060aca68-9071-46af-a627-ed80ad206d74\") " pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.365129 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x44pz\" (UniqueName: \"kubernetes.io/projected/f1c54538-1646-4a87-9240-fba80302af69-kube-api-access-x44pz\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.365197 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1c54538-1646-4a87-9240-fba80302af69-logs\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.365217 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.365254 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1c54538-1646-4a87-9240-fba80302af69-etc-machine-id\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.365331 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-scripts\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.365366 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.365403 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data-custom\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.365512 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1c54538-1646-4a87-9240-fba80302af69-etc-machine-id\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.366176 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1c54538-1646-4a87-9240-fba80302af69-logs\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.366705 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.369923 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-scripts\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.370678 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data-custom\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.372724 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.373365 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.381739 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x44pz\" (UniqueName: \"kubernetes.io/projected/f1c54538-1646-4a87-9240-fba80302af69-kube-api-access-x44pz\") pod \"manila-api-0\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.436322 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.881581 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:31:14 crc kubenswrapper[4787]: I1203 18:31:14.978321 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5767ddb7c-wsddg"] Dec 03 18:31:15 crc kubenswrapper[4787]: I1203 18:31:15.053660 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:31:15 crc kubenswrapper[4787]: I1203 18:31:15.155558 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 03 18:31:15 crc kubenswrapper[4787]: I1203 18:31:15.480520 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2a57671b-0fbe-437c-824b-5a6bdaf86f61","Type":"ContainerStarted","Data":"e45a7964b3e6dbf7f912168a94baccb97649b5736d9e7c6e1580e29d5993832c"} Dec 03 18:31:15 crc kubenswrapper[4787]: I1203 18:31:15.483267 4787 generic.go:334] "Generic (PLEG): container finished" podID="060aca68-9071-46af-a627-ed80ad206d74" containerID="da564b417c49dcba9101b7ad23c6c4fdd294c2848c1cfd9000d845d296271763" exitCode=0 Dec 03 18:31:15 crc kubenswrapper[4787]: I1203 18:31:15.483315 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" event={"ID":"060aca68-9071-46af-a627-ed80ad206d74","Type":"ContainerDied","Data":"da564b417c49dcba9101b7ad23c6c4fdd294c2848c1cfd9000d845d296271763"} Dec 03 18:31:15 crc kubenswrapper[4787]: I1203 18:31:15.483333 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" event={"ID":"060aca68-9071-46af-a627-ed80ad206d74","Type":"ContainerStarted","Data":"1b6e3f79b10637243d310b815cedfecf6a9b9f67cb0ac3b0bdc7e1adb45bbf8f"} Dec 03 18:31:15 crc kubenswrapper[4787]: I1203 18:31:15.485835 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f1c54538-1646-4a87-9240-fba80302af69","Type":"ContainerStarted","Data":"e4390c70012968e2a0c39dcfe79c409cd201c537b65318cb2090dc1a950c3098"} Dec 03 18:31:15 crc kubenswrapper[4787]: I1203 18:31:15.488366 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa","Type":"ContainerStarted","Data":"ef0f84737ff82258933a317fd5b7535c53f607fbd29e5525d8afcddbeaa4ae07"} Dec 03 18:31:16 crc kubenswrapper[4787]: I1203 18:31:16.502920 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" event={"ID":"060aca68-9071-46af-a627-ed80ad206d74","Type":"ContainerStarted","Data":"de94ae66a0d9a1b6f8445613eb5978908b8d3b1bb69cdd86132bef3e501b03c8"} Dec 03 18:31:16 crc kubenswrapper[4787]: I1203 18:31:16.503362 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:16 crc kubenswrapper[4787]: I1203 18:31:16.505698 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f1c54538-1646-4a87-9240-fba80302af69","Type":"ContainerStarted","Data":"6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b"} Dec 03 18:31:16 crc kubenswrapper[4787]: I1203 18:31:16.525201 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" podStartSLOduration=3.525186362 podStartE2EDuration="3.525186362s" podCreationTimestamp="2025-12-03 18:31:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:16.523083016 +0000 UTC m=+4693.340553985" watchObservedRunningTime="2025-12-03 18:31:16.525186362 +0000 UTC m=+4693.342657321" Dec 03 18:31:16 crc kubenswrapper[4787]: I1203 18:31:16.843171 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 03 18:31:17 crc kubenswrapper[4787]: I1203 18:31:17.528770 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f1c54538-1646-4a87-9240-fba80302af69","Type":"ContainerStarted","Data":"3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72"} Dec 03 18:31:17 crc kubenswrapper[4787]: I1203 18:31:17.530140 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 03 18:31:17 crc kubenswrapper[4787]: I1203 18:31:17.529474 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="f1c54538-1646-4a87-9240-fba80302af69" containerName="manila-api" containerID="cri-o://3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72" gracePeriod=30 Dec 03 18:31:17 crc kubenswrapper[4787]: I1203 18:31:17.528885 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="f1c54538-1646-4a87-9240-fba80302af69" containerName="manila-api-log" containerID="cri-o://6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b" gracePeriod=30 Dec 03 18:31:17 crc kubenswrapper[4787]: I1203 18:31:17.537911 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa","Type":"ContainerStarted","Data":"8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5"} Dec 03 18:31:17 crc kubenswrapper[4787]: I1203 18:31:17.566907 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=3.566886808 podStartE2EDuration="3.566886808s" podCreationTimestamp="2025-12-03 18:31:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:17.549792622 +0000 UTC m=+4694.367263601" watchObservedRunningTime="2025-12-03 18:31:17.566886808 +0000 UTC m=+4694.384357767" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.333456 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.525516 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x44pz\" (UniqueName: \"kubernetes.io/projected/f1c54538-1646-4a87-9240-fba80302af69-kube-api-access-x44pz\") pod \"f1c54538-1646-4a87-9240-fba80302af69\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.526438 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-scripts\") pod \"f1c54538-1646-4a87-9240-fba80302af69\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.526612 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1c54538-1646-4a87-9240-fba80302af69-logs\") pod \"f1c54538-1646-4a87-9240-fba80302af69\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.526640 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1c54538-1646-4a87-9240-fba80302af69-etc-machine-id\") pod \"f1c54538-1646-4a87-9240-fba80302af69\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.526703 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data\") pod \"f1c54538-1646-4a87-9240-fba80302af69\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.526752 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data-custom\") pod \"f1c54538-1646-4a87-9240-fba80302af69\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.526769 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-combined-ca-bundle\") pod \"f1c54538-1646-4a87-9240-fba80302af69\" (UID: \"f1c54538-1646-4a87-9240-fba80302af69\") " Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.526780 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f1c54538-1646-4a87-9240-fba80302af69-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f1c54538-1646-4a87-9240-fba80302af69" (UID: "f1c54538-1646-4a87-9240-fba80302af69"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.526982 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1c54538-1646-4a87-9240-fba80302af69-logs" (OuterVolumeSpecName: "logs") pod "f1c54538-1646-4a87-9240-fba80302af69" (UID: "f1c54538-1646-4a87-9240-fba80302af69"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.527432 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f1c54538-1646-4a87-9240-fba80302af69-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.527468 4787 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f1c54538-1646-4a87-9240-fba80302af69-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.553057 4787 generic.go:334] "Generic (PLEG): container finished" podID="f1c54538-1646-4a87-9240-fba80302af69" containerID="3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72" exitCode=0 Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.553089 4787 generic.go:334] "Generic (PLEG): container finished" podID="f1c54538-1646-4a87-9240-fba80302af69" containerID="6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b" exitCode=143 Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.553127 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f1c54538-1646-4a87-9240-fba80302af69","Type":"ContainerDied","Data":"3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72"} Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.553154 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f1c54538-1646-4a87-9240-fba80302af69","Type":"ContainerDied","Data":"6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b"} Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.553164 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"f1c54538-1646-4a87-9240-fba80302af69","Type":"ContainerDied","Data":"e4390c70012968e2a0c39dcfe79c409cd201c537b65318cb2090dc1a950c3098"} Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.553179 4787 scope.go:117] "RemoveContainer" containerID="3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.553531 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.558333 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa","Type":"ContainerStarted","Data":"48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294"} Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.592226 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=4.720752466 podStartE2EDuration="5.592209916s" podCreationTimestamp="2025-12-03 18:31:13 +0000 UTC" firstStartedPulling="2025-12-03 18:31:15.063234874 +0000 UTC m=+4691.880705833" lastFinishedPulling="2025-12-03 18:31:15.934692324 +0000 UTC m=+4692.752163283" observedRunningTime="2025-12-03 18:31:18.588415645 +0000 UTC m=+4695.405886624" watchObservedRunningTime="2025-12-03 18:31:18.592209916 +0000 UTC m=+4695.409680875" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.615624 4787 scope.go:117] "RemoveContainer" containerID="6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.653150 4787 scope.go:117] "RemoveContainer" containerID="3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72" Dec 03 18:31:18 crc kubenswrapper[4787]: E1203 18:31:18.653779 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72\": container with ID starting with 3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72 not found: ID does not exist" containerID="3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.653823 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72"} err="failed to get container status \"3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72\": rpc error: code = NotFound desc = could not find container \"3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72\": container with ID starting with 3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72 not found: ID does not exist" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.653850 4787 scope.go:117] "RemoveContainer" containerID="6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b" Dec 03 18:31:18 crc kubenswrapper[4787]: E1203 18:31:18.654218 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b\": container with ID starting with 6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b not found: ID does not exist" containerID="6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.654255 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b"} err="failed to get container status \"6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b\": rpc error: code = NotFound desc = could not find container \"6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b\": container with ID starting with 6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b not found: ID does not exist" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.654273 4787 scope.go:117] "RemoveContainer" containerID="3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.657319 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72"} err="failed to get container status \"3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72\": rpc error: code = NotFound desc = could not find container \"3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72\": container with ID starting with 3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72 not found: ID does not exist" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.657359 4787 scope.go:117] "RemoveContainer" containerID="6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.657862 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b"} err="failed to get container status \"6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b\": rpc error: code = NotFound desc = could not find container \"6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b\": container with ID starting with 6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b not found: ID does not exist" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.767800 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:31:18 crc kubenswrapper[4787]: I1203 18:31:18.814984 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.013892 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1c54538-1646-4a87-9240-fba80302af69-kube-api-access-x44pz" (OuterVolumeSpecName: "kube-api-access-x44pz") pod "f1c54538-1646-4a87-9240-fba80302af69" (UID: "f1c54538-1646-4a87-9240-fba80302af69"). InnerVolumeSpecName "kube-api-access-x44pz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.014065 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f1c54538-1646-4a87-9240-fba80302af69" (UID: "f1c54538-1646-4a87-9240-fba80302af69"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.027234 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-scripts" (OuterVolumeSpecName: "scripts") pod "f1c54538-1646-4a87-9240-fba80302af69" (UID: "f1c54538-1646-4a87-9240-fba80302af69"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.038263 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.038298 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x44pz\" (UniqueName: \"kubernetes.io/projected/f1c54538-1646-4a87-9240-fba80302af69-kube-api-access-x44pz\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.038311 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.064144 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f1c54538-1646-4a87-9240-fba80302af69" (UID: "f1c54538-1646-4a87-9240-fba80302af69"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.110049 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data" (OuterVolumeSpecName: "config-data") pod "f1c54538-1646-4a87-9240-fba80302af69" (UID: "f1c54538-1646-4a87-9240-fba80302af69"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.140898 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.140937 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f1c54538-1646-4a87-9240-fba80302af69-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.329313 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.341995 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.397346 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 03 18:31:19 crc kubenswrapper[4787]: E1203 18:31:19.397834 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1c54538-1646-4a87-9240-fba80302af69" containerName="manila-api" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.397849 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1c54538-1646-4a87-9240-fba80302af69" containerName="manila-api" Dec 03 18:31:19 crc kubenswrapper[4787]: E1203 18:31:19.397867 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1c54538-1646-4a87-9240-fba80302af69" containerName="manila-api-log" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.397873 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1c54538-1646-4a87-9240-fba80302af69" containerName="manila-api-log" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.398113 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1c54538-1646-4a87-9240-fba80302af69" containerName="manila-api-log" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.398147 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1c54538-1646-4a87-9240-fba80302af69" containerName="manila-api" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.399430 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.404570 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.404808 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.404962 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.428183 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.454501 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.454849 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="ceilometer-central-agent" containerID="cri-o://8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb" gracePeriod=30 Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.455128 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="proxy-httpd" containerID="cri-o://4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa" gracePeriod=30 Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.455317 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="ceilometer-notification-agent" containerID="cri-o://4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c" gracePeriod=30 Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.455359 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="sg-core" containerID="cri-o://db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9" gracePeriod=30 Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.556174 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-internal-tls-certs\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.556533 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01cfd003-6a74-4b88-a56d-0daa4cf79daf-logs\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.556557 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-config-data-custom\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.556633 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/01cfd003-6a74-4b88-a56d-0daa4cf79daf-etc-machine-id\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.556721 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.556784 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q8s8q\" (UniqueName: \"kubernetes.io/projected/01cfd003-6a74-4b88-a56d-0daa4cf79daf-kube-api-access-q8s8q\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.556852 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-scripts\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.556878 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-public-tls-certs\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.556914 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-config-data\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.659431 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.659509 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q8s8q\" (UniqueName: \"kubernetes.io/projected/01cfd003-6a74-4b88-a56d-0daa4cf79daf-kube-api-access-q8s8q\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.659565 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-scripts\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.659584 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-public-tls-certs\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.659609 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-config-data\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.659656 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-internal-tls-certs\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.659721 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01cfd003-6a74-4b88-a56d-0daa4cf79daf-logs\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.659741 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-config-data-custom\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.659794 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/01cfd003-6a74-4b88-a56d-0daa4cf79daf-etc-machine-id\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.659884 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/01cfd003-6a74-4b88-a56d-0daa4cf79daf-etc-machine-id\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.661061 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/01cfd003-6a74-4b88-a56d-0daa4cf79daf-logs\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.671982 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-config-data\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.673452 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-internal-tls-certs\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.677582 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-config-data-custom\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.678097 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-public-tls-certs\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.679921 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-scripts\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.679945 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/01cfd003-6a74-4b88-a56d-0daa4cf79daf-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.681477 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q8s8q\" (UniqueName: \"kubernetes.io/projected/01cfd003-6a74-4b88-a56d-0daa4cf79daf-kube-api-access-q8s8q\") pod \"manila-api-0\" (UID: \"01cfd003-6a74-4b88-a56d-0daa4cf79daf\") " pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.749555 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 03 18:31:19 crc kubenswrapper[4787]: I1203 18:31:19.780620 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1c54538-1646-4a87-9240-fba80302af69" path="/var/lib/kubelet/pods/f1c54538-1646-4a87-9240-fba80302af69/volumes" Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.421555 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.604566 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"01cfd003-6a74-4b88-a56d-0daa4cf79daf","Type":"ContainerStarted","Data":"ab00986577b286341862cf3a488fdcb7827bd267e6ee5895637c538c66c501a3"} Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.610790 4787 generic.go:334] "Generic (PLEG): container finished" podID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerID="4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa" exitCode=0 Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.610823 4787 generic.go:334] "Generic (PLEG): container finished" podID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerID="db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9" exitCode=2 Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.610831 4787 generic.go:334] "Generic (PLEG): container finished" podID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerID="8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb" exitCode=0 Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.610877 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c5f585a-69da-4588-ab44-bc08513f4c9f","Type":"ContainerDied","Data":"4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa"} Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.610901 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c5f585a-69da-4588-ab44-bc08513f4c9f","Type":"ContainerDied","Data":"db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9"} Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.610910 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c5f585a-69da-4588-ab44-bc08513f4c9f","Type":"ContainerDied","Data":"8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb"} Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.721090 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-5d686989bb-9689v" Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.805826 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66fb6d975b-t28ks"] Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.806108 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66fb6d975b-t28ks" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon-log" containerID="cri-o://422f921f2714b4435cd12b9578e5ee5f5f35831b82e8bae5717d3c307be5667b" gracePeriod=30 Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.806640 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-66fb6d975b-t28ks" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon" containerID="cri-o://4de5b41a3df3c3e88ba31d7984a50c7e78af565c2be91b7f608830527d0c964a" gracePeriod=30 Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.817270 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:31:20 crc kubenswrapper[4787]: I1203 18:31:20.831332 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-66fb6d975b-t28ks" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.71:8443/dashboard/auth/login/?next=/dashboard/\": read tcp 10.217.0.2:49210->10.217.1.71:8443: read: connection reset by peer" Dec 03 18:31:21 crc kubenswrapper[4787]: I1203 18:31:21.627154 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"01cfd003-6a74-4b88-a56d-0daa4cf79daf","Type":"ContainerStarted","Data":"049b33bb840459da65c6d783712c55ee2658bf7101ab9c26ef96834db130b5e5"} Dec 03 18:31:21 crc kubenswrapper[4787]: I1203 18:31:21.660601 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="proxy-httpd" probeResult="failure" output="Get \"https://10.217.0.253:3000/\": dial tcp 10.217.0.253:3000: connect: connection refused" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.201903 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.242733 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.308200 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-config-data\") pod \"1c5f585a-69da-4588-ab44-bc08513f4c9f\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.308310 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6w84c\" (UniqueName: \"kubernetes.io/projected/1c5f585a-69da-4588-ab44-bc08513f4c9f-kube-api-access-6w84c\") pod \"1c5f585a-69da-4588-ab44-bc08513f4c9f\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.308383 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-run-httpd\") pod \"1c5f585a-69da-4588-ab44-bc08513f4c9f\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.308425 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-ceilometer-tls-certs\") pod \"1c5f585a-69da-4588-ab44-bc08513f4c9f\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.308489 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-scripts\") pod \"1c5f585a-69da-4588-ab44-bc08513f4c9f\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.308595 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-sg-core-conf-yaml\") pod \"1c5f585a-69da-4588-ab44-bc08513f4c9f\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.308625 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-log-httpd\") pod \"1c5f585a-69da-4588-ab44-bc08513f4c9f\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.308768 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-combined-ca-bundle\") pod \"1c5f585a-69da-4588-ab44-bc08513f4c9f\" (UID: \"1c5f585a-69da-4588-ab44-bc08513f4c9f\") " Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.309119 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "1c5f585a-69da-4588-ab44-bc08513f4c9f" (UID: "1c5f585a-69da-4588-ab44-bc08513f4c9f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.309412 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.309564 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "1c5f585a-69da-4588-ab44-bc08513f4c9f" (UID: "1c5f585a-69da-4588-ab44-bc08513f4c9f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.315845 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c5f585a-69da-4588-ab44-bc08513f4c9f-kube-api-access-6w84c" (OuterVolumeSpecName: "kube-api-access-6w84c") pod "1c5f585a-69da-4588-ab44-bc08513f4c9f" (UID: "1c5f585a-69da-4588-ab44-bc08513f4c9f"). InnerVolumeSpecName "kube-api-access-6w84c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.321704 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-scripts" (OuterVolumeSpecName: "scripts") pod "1c5f585a-69da-4588-ab44-bc08513f4c9f" (UID: "1c5f585a-69da-4588-ab44-bc08513f4c9f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.348762 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "1c5f585a-69da-4588-ab44-bc08513f4c9f" (UID: "1c5f585a-69da-4588-ab44-bc08513f4c9f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.375370 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5767ddb7c-wsddg" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.409199 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "1c5f585a-69da-4588-ab44-bc08513f4c9f" (UID: "1c5f585a-69da-4588-ab44-bc08513f4c9f"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.413496 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6w84c\" (UniqueName: \"kubernetes.io/projected/1c5f585a-69da-4588-ab44-bc08513f4c9f-kube-api-access-6w84c\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.413538 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.413547 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.413556 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.413567 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/1c5f585a-69da-4588-ab44-bc08513f4c9f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.438000 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-hfmfg"] Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.438499 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" podUID="7e131b5d-ae30-4381-b9ec-b14feffb63ee" containerName="dnsmasq-dns" containerID="cri-o://130b047da394f7af9e253c5cd18ca0b26b05e55da8bafe1287b0c171a0fc939a" gracePeriod=10 Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.465744 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1c5f585a-69da-4588-ab44-bc08513f4c9f" (UID: "1c5f585a-69da-4588-ab44-bc08513f4c9f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.468621 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-config-data" (OuterVolumeSpecName: "config-data") pod "1c5f585a-69da-4588-ab44-bc08513f4c9f" (UID: "1c5f585a-69da-4588-ab44-bc08513f4c9f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.515597 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.515629 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1c5f585a-69da-4588-ab44-bc08513f4c9f-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.679564 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"01cfd003-6a74-4b88-a56d-0daa4cf79daf","Type":"ContainerStarted","Data":"ad24f65a5c1287b5be2999da86fc6611dfb8fa94fa6e2b8ec064d5324a481b22"} Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.681260 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.687971 4787 generic.go:334] "Generic (PLEG): container finished" podID="7e131b5d-ae30-4381-b9ec-b14feffb63ee" containerID="130b047da394f7af9e253c5cd18ca0b26b05e55da8bafe1287b0c171a0fc939a" exitCode=0 Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.688048 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" event={"ID":"7e131b5d-ae30-4381-b9ec-b14feffb63ee","Type":"ContainerDied","Data":"130b047da394f7af9e253c5cd18ca0b26b05e55da8bafe1287b0c171a0fc939a"} Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.691680 4787 generic.go:334] "Generic (PLEG): container finished" podID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerID="4de5b41a3df3c3e88ba31d7984a50c7e78af565c2be91b7f608830527d0c964a" exitCode=0 Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.691727 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fb6d975b-t28ks" event={"ID":"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3","Type":"ContainerDied","Data":"4de5b41a3df3c3e88ba31d7984a50c7e78af565c2be91b7f608830527d0c964a"} Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.696431 4787 generic.go:334] "Generic (PLEG): container finished" podID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerID="4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c" exitCode=0 Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.696488 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c5f585a-69da-4588-ab44-bc08513f4c9f","Type":"ContainerDied","Data":"4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c"} Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.696513 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"1c5f585a-69da-4588-ab44-bc08513f4c9f","Type":"ContainerDied","Data":"0827e4d2014acf3c689ae6906dfcb5a72c7c820a60270194027f058e645ee732"} Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.696529 4787 scope.go:117] "RemoveContainer" containerID="4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.696687 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.722146 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2a57671b-0fbe-437c-824b-5a6bdaf86f61","Type":"ContainerStarted","Data":"19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037"} Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.743934 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=5.743912373 podStartE2EDuration="5.743912373s" podCreationTimestamp="2025-12-03 18:31:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:24.711360603 +0000 UTC m=+4701.528831582" watchObservedRunningTime="2025-12-03 18:31:24.743912373 +0000 UTC m=+4701.561383332" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.765111 4787 scope.go:117] "RemoveContainer" containerID="db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.811908 4787 scope.go:117] "RemoveContainer" containerID="4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.813052 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.837368 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.850478 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:24 crc kubenswrapper[4787]: E1203 18:31:24.851555 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="ceilometer-central-agent" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.851593 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="ceilometer-central-agent" Dec 03 18:31:24 crc kubenswrapper[4787]: E1203 18:31:24.851628 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="sg-core" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.851635 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="sg-core" Dec 03 18:31:24 crc kubenswrapper[4787]: E1203 18:31:24.851646 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="ceilometer-notification-agent" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.851671 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="ceilometer-notification-agent" Dec 03 18:31:24 crc kubenswrapper[4787]: E1203 18:31:24.851704 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="proxy-httpd" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.851710 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="proxy-httpd" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.852507 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="ceilometer-notification-agent" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.852543 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="sg-core" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.852578 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="ceilometer-central-agent" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.852601 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" containerName="proxy-httpd" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.854938 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.858673 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.863042 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.863318 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.867403 4787 scope.go:117] "RemoveContainer" containerID="8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.887705 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.898286 4787 scope.go:117] "RemoveContainer" containerID="4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa" Dec 03 18:31:24 crc kubenswrapper[4787]: E1203 18:31:24.898729 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa\": container with ID starting with 4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa not found: ID does not exist" containerID="4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.898756 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa"} err="failed to get container status \"4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa\": rpc error: code = NotFound desc = could not find container \"4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa\": container with ID starting with 4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa not found: ID does not exist" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.898777 4787 scope.go:117] "RemoveContainer" containerID="db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9" Dec 03 18:31:24 crc kubenswrapper[4787]: E1203 18:31:24.898963 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9\": container with ID starting with db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9 not found: ID does not exist" containerID="db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.898983 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9"} err="failed to get container status \"db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9\": rpc error: code = NotFound desc = could not find container \"db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9\": container with ID starting with db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9 not found: ID does not exist" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.898994 4787 scope.go:117] "RemoveContainer" containerID="4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c" Dec 03 18:31:24 crc kubenswrapper[4787]: E1203 18:31:24.899207 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c\": container with ID starting with 4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c not found: ID does not exist" containerID="4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.899231 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c"} err="failed to get container status \"4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c\": rpc error: code = NotFound desc = could not find container \"4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c\": container with ID starting with 4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c not found: ID does not exist" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.899244 4787 scope.go:117] "RemoveContainer" containerID="8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb" Dec 03 18:31:24 crc kubenswrapper[4787]: E1203 18:31:24.899405 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb\": container with ID starting with 8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb not found: ID does not exist" containerID="8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb" Dec 03 18:31:24 crc kubenswrapper[4787]: I1203 18:31:24.899426 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb"} err="failed to get container status \"8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb\": rpc error: code = NotFound desc = could not find container \"8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb\": container with ID starting with 8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb not found: ID does not exist" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.034761 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-run-httpd\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.034819 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.034840 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-log-httpd\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.034856 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lszhk\" (UniqueName: \"kubernetes.io/projected/2e438905-3157-42a3-8e2f-3fa24a1b75c8-kube-api-access-lszhk\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.034899 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.034957 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-config-data\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.035004 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-scripts\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.035062 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.117458 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.136587 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-openstack-edpm-ipam\") pod \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.136635 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8hgb\" (UniqueName: \"kubernetes.io/projected/7e131b5d-ae30-4381-b9ec-b14feffb63ee-kube-api-access-s8hgb\") pod \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.136672 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-swift-storage-0\") pod \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.136811 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-sb\") pod \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.136843 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-config\") pod \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.136863 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-nb\") pod \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.136891 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-svc\") pod \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\" (UID: \"7e131b5d-ae30-4381-b9ec-b14feffb63ee\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.137009 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.137079 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-config-data\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.137133 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-scripts\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.137175 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.137255 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-run-httpd\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.137308 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.137327 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-log-httpd\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.137343 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lszhk\" (UniqueName: \"kubernetes.io/projected/2e438905-3157-42a3-8e2f-3fa24a1b75c8-kube-api-access-lszhk\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.142420 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-log-httpd\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.142675 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-run-httpd\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.157594 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.165635 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-config-data\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.166165 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.167745 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7e131b5d-ae30-4381-b9ec-b14feffb63ee-kube-api-access-s8hgb" (OuterVolumeSpecName: "kube-api-access-s8hgb") pod "7e131b5d-ae30-4381-b9ec-b14feffb63ee" (UID: "7e131b5d-ae30-4381-b9ec-b14feffb63ee"). InnerVolumeSpecName "kube-api-access-s8hgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.182670 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-scripts\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.191396 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: W1203 18:31:25.208430 4787 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1c54538_1646_4a87_9240_fba80302af69.slice/crio-conmon-3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1c54538_1646_4a87_9240_fba80302af69.slice/crio-conmon-3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72.scope: no such file or directory Dec 03 18:31:25 crc kubenswrapper[4787]: W1203 18:31:25.208547 4787 watcher.go:93] Error while processing event ("/sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1c54538_1646_4a87_9240_fba80302af69.slice/crio-3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1c54538_1646_4a87_9240_fba80302af69.slice/crio-3f08e5f01c3206837e745ef4f9f2942d40c8fa50088907c5a254ef5814319a72.scope: no such file or directory Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.218599 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lszhk\" (UniqueName: \"kubernetes.io/projected/2e438905-3157-42a3-8e2f-3fa24a1b75c8-kube-api-access-lszhk\") pod \"ceilometer-0\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: W1203 18:31:25.235702 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1c54538_1646_4a87_9240_fba80302af69.slice/crio-6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b.scope WatchSource:0}: Error finding container 6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b: Status 404 returned error can't find the container with id 6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.242981 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8hgb\" (UniqueName: \"kubernetes.io/projected/7e131b5d-ae30-4381-b9ec-b14feffb63ee-kube-api-access-s8hgb\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.301353 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "7e131b5d-ae30-4381-b9ec-b14feffb63ee" (UID: "7e131b5d-ae30-4381-b9ec-b14feffb63ee"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.325787 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7e131b5d-ae30-4381-b9ec-b14feffb63ee" (UID: "7e131b5d-ae30-4381-b9ec-b14feffb63ee"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.348219 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.348241 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.356575 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7e131b5d-ae30-4381-b9ec-b14feffb63ee" (UID: "7e131b5d-ae30-4381-b9ec-b14feffb63ee"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.386556 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7e131b5d-ae30-4381-b9ec-b14feffb63ee" (UID: "7e131b5d-ae30-4381-b9ec-b14feffb63ee"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.402948 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-config" (OuterVolumeSpecName: "config") pod "7e131b5d-ae30-4381-b9ec-b14feffb63ee" (UID: "7e131b5d-ae30-4381-b9ec-b14feffb63ee"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.420314 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7e131b5d-ae30-4381-b9ec-b14feffb63ee" (UID: "7e131b5d-ae30-4381-b9ec-b14feffb63ee"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.451738 4787 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-config\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.451776 4787 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.451787 4787 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.451798 4787 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7e131b5d-ae30-4381-b9ec-b14feffb63ee-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.496623 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.691284 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-66fb6d975b-t28ks" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.71:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.71:8443: connect: connection refused" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.742438 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" event={"ID":"7e131b5d-ae30-4381-b9ec-b14feffb63ee","Type":"ContainerDied","Data":"9db35b5bd5868f9d73eff33c7d83a47df7d2b02b371f2ea2feddc5dbe05e566d"} Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.742509 4787 scope.go:117] "RemoveContainer" containerID="130b047da394f7af9e253c5cd18ca0b26b05e55da8bafe1287b0c171a0fc939a" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.742647 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 18:31:25 crc kubenswrapper[4787]: E1203 18:31:25.767395 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c5f585a_69da_4588_ab44_bc08513f4c9f.slice/crio-4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c5f585a_69da_4588_ab44_bc08513f4c9f.slice/crio-db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00092e4b_4bb9_47bc_83e2_4fde7fe32313.slice/crio-50e484e7525c15bfd0b2aeede9f32e007d3d99a162172b8e29eb7dcea9553f17.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf57a600a_addf_458b_b146_8b294fb7fbc1.slice/crio-conmon-2f3b8125d69649225b1ee59bbdf1301d30207516a0d4fecc019b817167db5047.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1c54538_1646_4a87_9240_fba80302af69.slice/crio-e4390c70012968e2a0c39dcfe79c409cd201c537b65318cb2090dc1a950c3098\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00092e4b_4bb9_47bc_83e2_4fde7fe32313.slice/crio-d4095d3a681b531029bf766106b6018170fabda894ecb834035206ad413f2afc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1c54538_1646_4a87_9240_fba80302af69.slice/crio-conmon-6993eff134dc6011907c2867bb3e2656e142aff4f03dc3dfe83231acf839919b.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf57a600a_addf_458b_b146_8b294fb7fbc1.slice/crio-2f3b8125d69649225b1ee59bbdf1301d30207516a0d4fecc019b817167db5047.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e131b5d_ae30_4381_b9ec_b14feffb63ee.slice/crio-130b047da394f7af9e253c5cd18ca0b26b05e55da8bafe1287b0c171a0fc939a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf1c54538_1646_4a87_9240_fba80302af69.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c5f585a_69da_4588_ab44_bc08513f4c9f.slice/crio-0827e4d2014acf3c689ae6906dfcb5a72c7c820a60270194027f058e645ee732\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c5f585a_69da_4588_ab44_bc08513f4c9f.slice/crio-conmon-4ec6b6dff75aab3deb25496ecd30610c6e442dc05abb72dd0e3d37f14881fdaa.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacabd0b4_b5f4_46a1_b5d1_cfce6f076ff3.slice/crio-conmon-4de5b41a3df3c3e88ba31d7984a50c7e78af565c2be91b7f608830527d0c964a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c5f585a_69da_4588_ab44_bc08513f4c9f.slice/crio-8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf57a600a_addf_458b_b146_8b294fb7fbc1.slice/crio-conmon-f27aba4fd6dbbc751ecab3a1bd1df6ec5e5e441eb49b4abe6e28358a091ecb41.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00092e4b_4bb9_47bc_83e2_4fde7fe32313.slice/crio-conmon-d4095d3a681b531029bf766106b6018170fabda894ecb834035206ad413f2afc.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7e131b5d_ae30_4381_b9ec_b14feffb63ee.slice/crio-conmon-130b047da394f7af9e253c5cd18ca0b26b05e55da8bafe1287b0c171a0fc939a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacabd0b4_b5f4_46a1_b5d1_cfce6f076ff3.slice/crio-4de5b41a3df3c3e88ba31d7984a50c7e78af565c2be91b7f608830527d0c964a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c5f585a_69da_4588_ab44_bc08513f4c9f.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf57a600a_addf_458b_b146_8b294fb7fbc1.slice/crio-f27aba4fd6dbbc751ecab3a1bd1df6ec5e5e441eb49b4abe6e28358a091ecb41.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c5f585a_69da_4588_ab44_bc08513f4c9f.slice/crio-conmon-4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c5f585a_69da_4588_ab44_bc08513f4c9f.slice/crio-conmon-db56340883d72d64e30e17c295b4bc251519734bc6ec307ebd02d37ae1512ff9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00092e4b_4bb9_47bc_83e2_4fde7fe32313.slice/crio-conmon-50e484e7525c15bfd0b2aeede9f32e007d3d99a162172b8e29eb7dcea9553f17.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c5f585a_69da_4588_ab44_bc08513f4c9f.slice/crio-4414777fc236110ece9ff3bed9fc856445f0cf102a6e62e9693134280dd2696c.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1c5f585a_69da_4588_ab44_bc08513f4c9f.slice/crio-conmon-8b3c0119dcc444d8745ae6249418072018c554ac022d8e7d86645d9e63fc5bbb.scope\": RecentStats: unable to find data in memory cache]" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.777855 4787 generic.go:334] "Generic (PLEG): container finished" podID="f57a600a-addf-458b-b146-8b294fb7fbc1" containerID="f27aba4fd6dbbc751ecab3a1bd1df6ec5e5e441eb49b4abe6e28358a091ecb41" exitCode=137 Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.777890 4787 generic.go:334] "Generic (PLEG): container finished" podID="f57a600a-addf-458b-b146-8b294fb7fbc1" containerID="2f3b8125d69649225b1ee59bbdf1301d30207516a0d4fecc019b817167db5047" exitCode=137 Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.796451 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c5f585a-69da-4588-ab44-bc08513f4c9f" path="/var/lib/kubelet/pods/1c5f585a-69da-4588-ab44-bc08513f4c9f/volumes" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.797579 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7df97bdcbf-fgcfd" event={"ID":"f57a600a-addf-458b-b146-8b294fb7fbc1","Type":"ContainerDied","Data":"f27aba4fd6dbbc751ecab3a1bd1df6ec5e5e441eb49b4abe6e28358a091ecb41"} Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.797603 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7df97bdcbf-fgcfd" event={"ID":"f57a600a-addf-458b-b146-8b294fb7fbc1","Type":"ContainerDied","Data":"2f3b8125d69649225b1ee59bbdf1301d30207516a0d4fecc019b817167db5047"} Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.822100 4787 generic.go:334] "Generic (PLEG): container finished" podID="00092e4b-4bb9-47bc-83e2-4fde7fe32313" containerID="d4095d3a681b531029bf766106b6018170fabda894ecb834035206ad413f2afc" exitCode=137 Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.822131 4787 generic.go:334] "Generic (PLEG): container finished" podID="00092e4b-4bb9-47bc-83e2-4fde7fe32313" containerID="50e484e7525c15bfd0b2aeede9f32e007d3d99a162172b8e29eb7dcea9553f17" exitCode=137 Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.822186 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c7b8ff54f-b242w" event={"ID":"00092e4b-4bb9-47bc-83e2-4fde7fe32313","Type":"ContainerDied","Data":"d4095d3a681b531029bf766106b6018170fabda894ecb834035206ad413f2afc"} Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.822357 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c7b8ff54f-b242w" event={"ID":"00092e4b-4bb9-47bc-83e2-4fde7fe32313","Type":"ContainerDied","Data":"50e484e7525c15bfd0b2aeede9f32e007d3d99a162172b8e29eb7dcea9553f17"} Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.824931 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2a57671b-0fbe-437c-824b-5a6bdaf86f61","Type":"ContainerStarted","Data":"6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012"} Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.849867 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=4.03667778 podStartE2EDuration="12.849851723s" podCreationTimestamp="2025-12-03 18:31:13 +0000 UTC" firstStartedPulling="2025-12-03 18:31:14.882229711 +0000 UTC m=+4691.699700670" lastFinishedPulling="2025-12-03 18:31:23.695403654 +0000 UTC m=+4700.512874613" observedRunningTime="2025-12-03 18:31:25.848461156 +0000 UTC m=+4702.665932115" watchObservedRunningTime="2025-12-03 18:31:25.849851723 +0000 UTC m=+4702.667322692" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.853789 4787 scope.go:117] "RemoveContainer" containerID="f8c1c616cbf2ebb0cdde7dabec1e2235a993486bc8f294188f8c9c14ba7b39a0" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.855795 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.965220 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nf8d9\" (UniqueName: \"kubernetes.io/projected/f57a600a-addf-458b-b146-8b294fb7fbc1-kube-api-access-nf8d9\") pod \"f57a600a-addf-458b-b146-8b294fb7fbc1\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.965284 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-config-data\") pod \"f57a600a-addf-458b-b146-8b294fb7fbc1\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.965324 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-scripts\") pod \"f57a600a-addf-458b-b146-8b294fb7fbc1\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.965386 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f57a600a-addf-458b-b146-8b294fb7fbc1-horizon-secret-key\") pod \"f57a600a-addf-458b-b146-8b294fb7fbc1\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.965565 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f57a600a-addf-458b-b146-8b294fb7fbc1-logs\") pod \"f57a600a-addf-458b-b146-8b294fb7fbc1\" (UID: \"f57a600a-addf-458b-b146-8b294fb7fbc1\") " Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.968389 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f57a600a-addf-458b-b146-8b294fb7fbc1-logs" (OuterVolumeSpecName: "logs") pod "f57a600a-addf-458b-b146-8b294fb7fbc1" (UID: "f57a600a-addf-458b-b146-8b294fb7fbc1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.973724 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f57a600a-addf-458b-b146-8b294fb7fbc1-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "f57a600a-addf-458b-b146-8b294fb7fbc1" (UID: "f57a600a-addf-458b-b146-8b294fb7fbc1"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.973778 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f57a600a-addf-458b-b146-8b294fb7fbc1-kube-api-access-nf8d9" (OuterVolumeSpecName: "kube-api-access-nf8d9") pod "f57a600a-addf-458b-b146-8b294fb7fbc1" (UID: "f57a600a-addf-458b-b146-8b294fb7fbc1"). InnerVolumeSpecName "kube-api-access-nf8d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:25 crc kubenswrapper[4787]: I1203 18:31:25.999913 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.002172 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-scripts" (OuterVolumeSpecName: "scripts") pod "f57a600a-addf-458b-b146-8b294fb7fbc1" (UID: "f57a600a-addf-458b-b146-8b294fb7fbc1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.050823 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-config-data" (OuterVolumeSpecName: "config-data") pod "f57a600a-addf-458b-b146-8b294fb7fbc1" (UID: "f57a600a-addf-458b-b146-8b294fb7fbc1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.069816 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.069852 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/f57a600a-addf-458b-b146-8b294fb7fbc1-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.069862 4787 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/f57a600a-addf-458b-b146-8b294fb7fbc1-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.069873 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f57a600a-addf-458b-b146-8b294fb7fbc1-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.069881 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nf8d9\" (UniqueName: \"kubernetes.io/projected/f57a600a-addf-458b-b146-8b294fb7fbc1-kube-api-access-nf8d9\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.097843 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.171353 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00092e4b-4bb9-47bc-83e2-4fde7fe32313-horizon-secret-key\") pod \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.171536 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-config-data\") pod \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.171564 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svbrp\" (UniqueName: \"kubernetes.io/projected/00092e4b-4bb9-47bc-83e2-4fde7fe32313-kube-api-access-svbrp\") pod \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.171588 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-scripts\") pod \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.171608 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00092e4b-4bb9-47bc-83e2-4fde7fe32313-logs\") pod \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\" (UID: \"00092e4b-4bb9-47bc-83e2-4fde7fe32313\") " Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.172906 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/00092e4b-4bb9-47bc-83e2-4fde7fe32313-logs" (OuterVolumeSpecName: "logs") pod "00092e4b-4bb9-47bc-83e2-4fde7fe32313" (UID: "00092e4b-4bb9-47bc-83e2-4fde7fe32313"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.176395 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00092e4b-4bb9-47bc-83e2-4fde7fe32313-kube-api-access-svbrp" (OuterVolumeSpecName: "kube-api-access-svbrp") pod "00092e4b-4bb9-47bc-83e2-4fde7fe32313" (UID: "00092e4b-4bb9-47bc-83e2-4fde7fe32313"). InnerVolumeSpecName "kube-api-access-svbrp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.176403 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00092e4b-4bb9-47bc-83e2-4fde7fe32313-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "00092e4b-4bb9-47bc-83e2-4fde7fe32313" (UID: "00092e4b-4bb9-47bc-83e2-4fde7fe32313"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.202118 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-config-data" (OuterVolumeSpecName: "config-data") pod "00092e4b-4bb9-47bc-83e2-4fde7fe32313" (UID: "00092e4b-4bb9-47bc-83e2-4fde7fe32313"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.203806 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-scripts" (OuterVolumeSpecName: "scripts") pod "00092e4b-4bb9-47bc-83e2-4fde7fe32313" (UID: "00092e4b-4bb9-47bc-83e2-4fde7fe32313"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.275924 4787 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/00092e4b-4bb9-47bc-83e2-4fde7fe32313-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.277331 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.277404 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svbrp\" (UniqueName: \"kubernetes.io/projected/00092e4b-4bb9-47bc-83e2-4fde7fe32313-kube-api-access-svbrp\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.277483 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/00092e4b-4bb9-47bc-83e2-4fde7fe32313-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:26 crc kubenswrapper[4787]: I1203 18:31:26.277554 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/00092e4b-4bb9-47bc-83e2-4fde7fe32313-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:26.848848 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-7df97bdcbf-fgcfd" event={"ID":"f57a600a-addf-458b-b146-8b294fb7fbc1","Type":"ContainerDied","Data":"3ddcd9aa090b310667987b521f90eb53d3161782cc265192431e141f120128e8"} Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:26.849140 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-7df97bdcbf-fgcfd" Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:26.849197 4787 scope.go:117] "RemoveContainer" containerID="f27aba4fd6dbbc751ecab3a1bd1df6ec5e5e441eb49b4abe6e28358a091ecb41" Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:26.851072 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e438905-3157-42a3-8e2f-3fa24a1b75c8","Type":"ContainerStarted","Data":"44090e1deec9d67815c478077aa3038893e46eb258ed036a6788c45da9d41c5b"} Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:26.851111 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e438905-3157-42a3-8e2f-3fa24a1b75c8","Type":"ContainerStarted","Data":"bc9f94467f62d8b1a6050bfe114d4293fa95cb247b76a63f530d04227edc652e"} Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:26.853247 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6c7b8ff54f-b242w" event={"ID":"00092e4b-4bb9-47bc-83e2-4fde7fe32313","Type":"ContainerDied","Data":"c3652d7b5ddf5183448c7a5bc204ba7c2df30e23c40062ca02408e212e846863"} Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:26.853319 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6c7b8ff54f-b242w" Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:26.945352 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-7df97bdcbf-fgcfd"] Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:26.959417 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-7df97bdcbf-fgcfd"] Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:26.969370 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6c7b8ff54f-b242w"] Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:26.984751 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6c7b8ff54f-b242w"] Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:27.054841 4787 scope.go:117] "RemoveContainer" containerID="2f3b8125d69649225b1ee59bbdf1301d30207516a0d4fecc019b817167db5047" Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:27.077720 4787 scope.go:117] "RemoveContainer" containerID="d4095d3a681b531029bf766106b6018170fabda894ecb834035206ad413f2afc" Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:27.257933 4787 scope.go:117] "RemoveContainer" containerID="50e484e7525c15bfd0b2aeede9f32e007d3d99a162172b8e29eb7dcea9553f17" Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:27.766408 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:31:27 crc kubenswrapper[4787]: E1203 18:31:27.767131 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:27.782357 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00092e4b-4bb9-47bc-83e2-4fde7fe32313" path="/var/lib/kubelet/pods/00092e4b-4bb9-47bc-83e2-4fde7fe32313/volumes" Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:27.783522 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f57a600a-addf-458b-b146-8b294fb7fbc1" path="/var/lib/kubelet/pods/f57a600a-addf-458b-b146-8b294fb7fbc1/volumes" Dec 03 18:31:27 crc kubenswrapper[4787]: I1203 18:31:27.866933 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e438905-3157-42a3-8e2f-3fa24a1b75c8","Type":"ContainerStarted","Data":"2e213c081473b6f716a628245ed15b68bbc0a9d5d80322447926689d2b594b05"} Dec 03 18:31:28 crc kubenswrapper[4787]: I1203 18:31:28.239280 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:29 crc kubenswrapper[4787]: I1203 18:31:29.892134 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e438905-3157-42a3-8e2f-3fa24a1b75c8","Type":"ContainerStarted","Data":"db9f2bebed810a317367c02a17c64f640ced191f4a650a3b4061021e13ea4f8e"} Dec 03 18:31:30 crc kubenswrapper[4787]: I1203 18:31:30.909085 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e438905-3157-42a3-8e2f-3fa24a1b75c8","Type":"ContainerStarted","Data":"ddcb4b3a16d83bcc83aa9e5774cd76a4d8a1a7ce755a77d866dd41d9ff9b5e20"} Dec 03 18:31:30 crc kubenswrapper[4787]: I1203 18:31:30.909369 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="ceilometer-central-agent" containerID="cri-o://44090e1deec9d67815c478077aa3038893e46eb258ed036a6788c45da9d41c5b" gracePeriod=30 Dec 03 18:31:30 crc kubenswrapper[4787]: I1203 18:31:30.912374 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 18:31:30 crc kubenswrapper[4787]: I1203 18:31:30.909412 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="proxy-httpd" containerID="cri-o://ddcb4b3a16d83bcc83aa9e5774cd76a4d8a1a7ce755a77d866dd41d9ff9b5e20" gracePeriod=30 Dec 03 18:31:30 crc kubenswrapper[4787]: I1203 18:31:30.909415 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="ceilometer-notification-agent" containerID="cri-o://2e213c081473b6f716a628245ed15b68bbc0a9d5d80322447926689d2b594b05" gracePeriod=30 Dec 03 18:31:30 crc kubenswrapper[4787]: I1203 18:31:30.909427 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="sg-core" containerID="cri-o://db9f2bebed810a317367c02a17c64f640ced191f4a650a3b4061021e13ea4f8e" gracePeriod=30 Dec 03 18:31:30 crc kubenswrapper[4787]: I1203 18:31:30.938932 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.767463236 podStartE2EDuration="6.938915784s" podCreationTimestamp="2025-12-03 18:31:24 +0000 UTC" firstStartedPulling="2025-12-03 18:31:26.096444448 +0000 UTC m=+4702.913915407" lastFinishedPulling="2025-12-03 18:31:30.267896996 +0000 UTC m=+4707.085367955" observedRunningTime="2025-12-03 18:31:30.932592525 +0000 UTC m=+4707.750063484" watchObservedRunningTime="2025-12-03 18:31:30.938915784 +0000 UTC m=+4707.756386743" Dec 03 18:31:31 crc kubenswrapper[4787]: I1203 18:31:31.926765 4787 generic.go:334] "Generic (PLEG): container finished" podID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerID="ddcb4b3a16d83bcc83aa9e5774cd76a4d8a1a7ce755a77d866dd41d9ff9b5e20" exitCode=0 Dec 03 18:31:31 crc kubenswrapper[4787]: I1203 18:31:31.927179 4787 generic.go:334] "Generic (PLEG): container finished" podID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerID="db9f2bebed810a317367c02a17c64f640ced191f4a650a3b4061021e13ea4f8e" exitCode=2 Dec 03 18:31:31 crc kubenswrapper[4787]: I1203 18:31:31.926830 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e438905-3157-42a3-8e2f-3fa24a1b75c8","Type":"ContainerDied","Data":"ddcb4b3a16d83bcc83aa9e5774cd76a4d8a1a7ce755a77d866dd41d9ff9b5e20"} Dec 03 18:31:31 crc kubenswrapper[4787]: I1203 18:31:31.927232 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e438905-3157-42a3-8e2f-3fa24a1b75c8","Type":"ContainerDied","Data":"db9f2bebed810a317367c02a17c64f640ced191f4a650a3b4061021e13ea4f8e"} Dec 03 18:31:31 crc kubenswrapper[4787]: I1203 18:31:31.927247 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e438905-3157-42a3-8e2f-3fa24a1b75c8","Type":"ContainerDied","Data":"2e213c081473b6f716a628245ed15b68bbc0a9d5d80322447926689d2b594b05"} Dec 03 18:31:31 crc kubenswrapper[4787]: I1203 18:31:31.927201 4787 generic.go:334] "Generic (PLEG): container finished" podID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerID="2e213c081473b6f716a628245ed15b68bbc0a9d5d80322447926689d2b594b05" exitCode=0 Dec 03 18:31:34 crc kubenswrapper[4787]: I1203 18:31:34.214196 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 03 18:31:34 crc kubenswrapper[4787]: I1203 18:31:34.967558 4787 generic.go:334] "Generic (PLEG): container finished" podID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerID="44090e1deec9d67815c478077aa3038893e46eb258ed036a6788c45da9d41c5b" exitCode=0 Dec 03 18:31:34 crc kubenswrapper[4787]: I1203 18:31:34.967741 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e438905-3157-42a3-8e2f-3fa24a1b75c8","Type":"ContainerDied","Data":"44090e1deec9d67815c478077aa3038893e46eb258ed036a6788c45da9d41c5b"} Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.691325 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-66fb6d975b-t28ks" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.71:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.71:8443: connect: connection refused" Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.691801 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.701092 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.801179 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.813360 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.870585 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.979257 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e438905-3157-42a3-8e2f-3fa24a1b75c8","Type":"ContainerDied","Data":"bc9f94467f62d8b1a6050bfe114d4293fa95cb247b76a63f530d04227edc652e"} Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.979513 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc9f94467f62d8b1a6050bfe114d4293fa95cb247b76a63f530d04227edc652e" Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.979627 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" containerName="probe" containerID="cri-o://48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294" gracePeriod=30 Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.979396 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" containerName="manila-scheduler" containerID="cri-o://8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5" gracePeriod=30 Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.979873 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="2a57671b-0fbe-437c-824b-5a6bdaf86f61" containerName="probe" containerID="cri-o://6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012" gracePeriod=30 Dec 03 18:31:35 crc kubenswrapper[4787]: I1203 18:31:35.979483 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="2a57671b-0fbe-437c-824b-5a6bdaf86f61" containerName="manila-share" containerID="cri-o://19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037" gracePeriod=30 Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.181194 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.320393 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-run-httpd\") pod \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.320485 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-log-httpd\") pod \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.320531 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lszhk\" (UniqueName: \"kubernetes.io/projected/2e438905-3157-42a3-8e2f-3fa24a1b75c8-kube-api-access-lszhk\") pod \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.320561 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-ceilometer-tls-certs\") pod \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.320639 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-sg-core-conf-yaml\") pod \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.320664 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-scripts\") pod \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.320689 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-config-data\") pod \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.320721 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-combined-ca-bundle\") pod \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\" (UID: \"2e438905-3157-42a3-8e2f-3fa24a1b75c8\") " Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.322452 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2e438905-3157-42a3-8e2f-3fa24a1b75c8" (UID: "2e438905-3157-42a3-8e2f-3fa24a1b75c8"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.322781 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2e438905-3157-42a3-8e2f-3fa24a1b75c8" (UID: "2e438905-3157-42a3-8e2f-3fa24a1b75c8"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.327757 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e438905-3157-42a3-8e2f-3fa24a1b75c8-kube-api-access-lszhk" (OuterVolumeSpecName: "kube-api-access-lszhk") pod "2e438905-3157-42a3-8e2f-3fa24a1b75c8" (UID: "2e438905-3157-42a3-8e2f-3fa24a1b75c8"). InnerVolumeSpecName "kube-api-access-lszhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.328340 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-scripts" (OuterVolumeSpecName: "scripts") pod "2e438905-3157-42a3-8e2f-3fa24a1b75c8" (UID: "2e438905-3157-42a3-8e2f-3fa24a1b75c8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.354497 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2e438905-3157-42a3-8e2f-3fa24a1b75c8" (UID: "2e438905-3157-42a3-8e2f-3fa24a1b75c8"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.408239 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "2e438905-3157-42a3-8e2f-3fa24a1b75c8" (UID: "2e438905-3157-42a3-8e2f-3fa24a1b75c8"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.426908 4787 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.426946 4787 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e438905-3157-42a3-8e2f-3fa24a1b75c8-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.426955 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lszhk\" (UniqueName: \"kubernetes.io/projected/2e438905-3157-42a3-8e2f-3fa24a1b75c8-kube-api-access-lszhk\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.426965 4787 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.426975 4787 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.426984 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.443442 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e438905-3157-42a3-8e2f-3fa24a1b75c8" (UID: "2e438905-3157-42a3-8e2f-3fa24a1b75c8"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.453681 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-config-data" (OuterVolumeSpecName: "config-data") pod "2e438905-3157-42a3-8e2f-3fa24a1b75c8" (UID: "2e438905-3157-42a3-8e2f-3fa24a1b75c8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.529161 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.529242 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e438905-3157-42a3-8e2f-3fa24a1b75c8-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.875609 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.990386 4787 generic.go:334] "Generic (PLEG): container finished" podID="2a57671b-0fbe-437c-824b-5a6bdaf86f61" containerID="6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012" exitCode=0 Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.990416 4787 generic.go:334] "Generic (PLEG): container finished" podID="2a57671b-0fbe-437c-824b-5a6bdaf86f61" containerID="19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037" exitCode=1 Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.990429 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.990469 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2a57671b-0fbe-437c-824b-5a6bdaf86f61","Type":"ContainerDied","Data":"6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012"} Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.990510 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2a57671b-0fbe-437c-824b-5a6bdaf86f61","Type":"ContainerDied","Data":"19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037"} Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.990521 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2a57671b-0fbe-437c-824b-5a6bdaf86f61","Type":"ContainerDied","Data":"e45a7964b3e6dbf7f912168a94baccb97649b5736d9e7c6e1580e29d5993832c"} Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.990539 4787 scope.go:117] "RemoveContainer" containerID="6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.993180 4787 generic.go:334] "Generic (PLEG): container finished" podID="27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" containerID="48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294" exitCode=0 Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.993296 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:31:36 crc kubenswrapper[4787]: I1203 18:31:36.993261 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa","Type":"ContainerDied","Data":"48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294"} Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.015530 4787 scope.go:117] "RemoveContainer" containerID="19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.033008 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.040870 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data\") pod \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.040930 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-etc-machine-id\") pod \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.041041 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data-custom\") pod \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.041159 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-combined-ca-bundle\") pod \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.041212 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wsvh8\" (UniqueName: \"kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-kube-api-access-wsvh8\") pod \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.041249 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-ceph\") pod \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.041319 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-var-lib-manila\") pod \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.041380 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-scripts\") pod \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\" (UID: \"2a57671b-0fbe-437c-824b-5a6bdaf86f61\") " Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.044295 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "2a57671b-0fbe-437c-824b-5a6bdaf86f61" (UID: "2a57671b-0fbe-437c-824b-5a6bdaf86f61"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.044682 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.044769 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2a57671b-0fbe-437c-824b-5a6bdaf86f61" (UID: "2a57671b-0fbe-437c-824b-5a6bdaf86f61"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.047403 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-ceph" (OuterVolumeSpecName: "ceph") pod "2a57671b-0fbe-437c-824b-5a6bdaf86f61" (UID: "2a57671b-0fbe-437c-824b-5a6bdaf86f61"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.047398 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-scripts" (OuterVolumeSpecName: "scripts") pod "2a57671b-0fbe-437c-824b-5a6bdaf86f61" (UID: "2a57671b-0fbe-437c-824b-5a6bdaf86f61"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.055873 4787 scope.go:117] "RemoveContainer" containerID="6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.059536 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-kube-api-access-wsvh8" (OuterVolumeSpecName: "kube-api-access-wsvh8") pod "2a57671b-0fbe-437c-824b-5a6bdaf86f61" (UID: "2a57671b-0fbe-437c-824b-5a6bdaf86f61"). InnerVolumeSpecName "kube-api-access-wsvh8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.059562 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012\": container with ID starting with 6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012 not found: ID does not exist" containerID="6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.059607 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012"} err="failed to get container status \"6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012\": rpc error: code = NotFound desc = could not find container \"6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012\": container with ID starting with 6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012 not found: ID does not exist" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.059633 4787 scope.go:117] "RemoveContainer" containerID="19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.059813 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2a57671b-0fbe-437c-824b-5a6bdaf86f61" (UID: "2a57671b-0fbe-437c-824b-5a6bdaf86f61"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.060261 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037\": container with ID starting with 19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037 not found: ID does not exist" containerID="19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.060291 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037"} err="failed to get container status \"19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037\": rpc error: code = NotFound desc = could not find container \"19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037\": container with ID starting with 19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037 not found: ID does not exist" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.060304 4787 scope.go:117] "RemoveContainer" containerID="6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.060656 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012"} err="failed to get container status \"6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012\": rpc error: code = NotFound desc = could not find container \"6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012\": container with ID starting with 6b883c94b77ae50361bda9017cd34289d821de3852e7813d4a152f3eebbf4012 not found: ID does not exist" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.060702 4787 scope.go:117] "RemoveContainer" containerID="19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.061002 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037"} err="failed to get container status \"19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037\": rpc error: code = NotFound desc = could not find container \"19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037\": container with ID starting with 19da2b05f538f25d3b2c8f559554559782fcb74d5e86038fa6c6d0124d887037 not found: ID does not exist" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.065081 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.066187 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00092e4b-4bb9-47bc-83e2-4fde7fe32313" containerName="horizon-log" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.066278 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="00092e4b-4bb9-47bc-83e2-4fde7fe32313" containerName="horizon-log" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.066297 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="proxy-httpd" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.066305 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="proxy-httpd" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.066359 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e131b5d-ae30-4381-b9ec-b14feffb63ee" containerName="dnsmasq-dns" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.066366 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e131b5d-ae30-4381-b9ec-b14feffb63ee" containerName="dnsmasq-dns" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.066380 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="ceilometer-central-agent" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.066387 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="ceilometer-central-agent" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.066425 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f57a600a-addf-458b-b146-8b294fb7fbc1" containerName="horizon-log" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.066432 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f57a600a-addf-458b-b146-8b294fb7fbc1" containerName="horizon-log" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.066447 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f57a600a-addf-458b-b146-8b294fb7fbc1" containerName="horizon" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.066454 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f57a600a-addf-458b-b146-8b294fb7fbc1" containerName="horizon" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.066468 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00092e4b-4bb9-47bc-83e2-4fde7fe32313" containerName="horizon" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.066510 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="00092e4b-4bb9-47bc-83e2-4fde7fe32313" containerName="horizon" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.066524 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7e131b5d-ae30-4381-b9ec-b14feffb63ee" containerName="init" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.066532 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7e131b5d-ae30-4381-b9ec-b14feffb63ee" containerName="init" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.066552 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a57671b-0fbe-437c-824b-5a6bdaf86f61" containerName="probe" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.066618 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a57671b-0fbe-437c-824b-5a6bdaf86f61" containerName="probe" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.066635 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a57671b-0fbe-437c-824b-5a6bdaf86f61" containerName="manila-share" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068088 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a57671b-0fbe-437c-824b-5a6bdaf86f61" containerName="manila-share" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.068113 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="ceilometer-notification-agent" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068135 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="ceilometer-notification-agent" Dec 03 18:31:37 crc kubenswrapper[4787]: E1203 18:31:37.068151 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="sg-core" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068156 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="sg-core" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068548 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="ceilometer-notification-agent" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068568 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="ceilometer-central-agent" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068579 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="00092e4b-4bb9-47bc-83e2-4fde7fe32313" containerName="horizon" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068591 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f57a600a-addf-458b-b146-8b294fb7fbc1" containerName="horizon" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068601 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="00092e4b-4bb9-47bc-83e2-4fde7fe32313" containerName="horizon-log" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068612 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7e131b5d-ae30-4381-b9ec-b14feffb63ee" containerName="dnsmasq-dns" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068626 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a57671b-0fbe-437c-824b-5a6bdaf86f61" containerName="manila-share" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068643 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="sg-core" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068650 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a57671b-0fbe-437c-824b-5a6bdaf86f61" containerName="probe" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068660 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" containerName="proxy-httpd" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.068670 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f57a600a-addf-458b-b146-8b294fb7fbc1" containerName="horizon-log" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.070921 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.073428 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.073681 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.074109 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.084584 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.143600 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wsvh8\" (UniqueName: \"kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-kube-api-access-wsvh8\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.143635 4787 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2a57671b-0fbe-437c-824b-5a6bdaf86f61-ceph\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.143645 4787 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-var-lib-manila\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.143654 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.143662 4787 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2a57671b-0fbe-437c-824b-5a6bdaf86f61-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.143670 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.208860 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2a57671b-0fbe-437c-824b-5a6bdaf86f61" (UID: "2a57671b-0fbe-437c-824b-5a6bdaf86f61"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.245230 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.245274 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-config-data\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.245309 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.245595 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11460db3-3d3a-426a-9980-f1dd41a84497-run-httpd\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.245674 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-82dkb\" (UniqueName: \"kubernetes.io/projected/11460db3-3d3a-426a-9980-f1dd41a84497-kube-api-access-82dkb\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.245707 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.245733 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-scripts\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.245844 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11460db3-3d3a-426a-9980-f1dd41a84497-log-httpd\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.246606 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.293460 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data" (OuterVolumeSpecName: "config-data") pod "2a57671b-0fbe-437c-824b-5a6bdaf86f61" (UID: "2a57671b-0fbe-437c-824b-5a6bdaf86f61"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.348444 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.348495 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11460db3-3d3a-426a-9980-f1dd41a84497-run-httpd\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.348547 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-82dkb\" (UniqueName: \"kubernetes.io/projected/11460db3-3d3a-426a-9980-f1dd41a84497-kube-api-access-82dkb\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.348569 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.348590 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-scripts\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.348616 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11460db3-3d3a-426a-9980-f1dd41a84497-log-httpd\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.348761 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.348780 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-config-data\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.348836 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a57671b-0fbe-437c-824b-5a6bdaf86f61-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.349279 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11460db3-3d3a-426a-9980-f1dd41a84497-run-httpd\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.349293 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/11460db3-3d3a-426a-9980-f1dd41a84497-log-httpd\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.353607 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-scripts\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.353770 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.353971 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.354436 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-config-data\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.355539 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11460db3-3d3a-426a-9980-f1dd41a84497-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.365344 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-82dkb\" (UniqueName: \"kubernetes.io/projected/11460db3-3d3a-426a-9980-f1dd41a84497-kube-api-access-82dkb\") pod \"ceilometer-0\" (UID: \"11460db3-3d3a-426a-9980-f1dd41a84497\") " pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.412891 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.661875 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.681702 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.706256 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.716247 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.719658 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.723226 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.782597 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a57671b-0fbe-437c-824b-5a6bdaf86f61" path="/var/lib/kubelet/pods/2a57671b-0fbe-437c-824b-5a6bdaf86f61/volumes" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.784237 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e438905-3157-42a3-8e2f-3fa24a1b75c8" path="/var/lib/kubelet/pods/2e438905-3157-42a3-8e2f-3fa24a1b75c8/volumes" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.864319 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-scripts\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.864371 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b52259b1-98e9-4791-a981-fc62fdd0e138-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.864467 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b52259b1-98e9-4791-a981-fc62fdd0e138-ceph\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.864497 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b52259b1-98e9-4791-a981-fc62fdd0e138-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.864511 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.864544 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjx57\" (UniqueName: \"kubernetes.io/projected/b52259b1-98e9-4791-a981-fc62fdd0e138-kube-api-access-xjx57\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.864666 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.864698 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-config-data\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.948346 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.966754 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-scripts\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.966815 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b52259b1-98e9-4791-a981-fc62fdd0e138-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.966955 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b52259b1-98e9-4791-a981-fc62fdd0e138-ceph\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.966990 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/b52259b1-98e9-4791-a981-fc62fdd0e138-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.967698 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b52259b1-98e9-4791-a981-fc62fdd0e138-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.967727 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.967741 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/b52259b1-98e9-4791-a981-fc62fdd0e138-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.967757 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjx57\" (UniqueName: \"kubernetes.io/projected/b52259b1-98e9-4791-a981-fc62fdd0e138-kube-api-access-xjx57\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.967878 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:37 crc kubenswrapper[4787]: I1203 18:31:37.967909 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-config-data\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:38 crc kubenswrapper[4787]: I1203 18:31:38.216854 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:38 crc kubenswrapper[4787]: I1203 18:31:38.216954 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-scripts\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:38 crc kubenswrapper[4787]: I1203 18:31:38.217512 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-config-data\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:38 crc kubenswrapper[4787]: I1203 18:31:38.219513 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b52259b1-98e9-4791-a981-fc62fdd0e138-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:38 crc kubenswrapper[4787]: I1203 18:31:38.220085 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjx57\" (UniqueName: \"kubernetes.io/projected/b52259b1-98e9-4791-a981-fc62fdd0e138-kube-api-access-xjx57\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:38 crc kubenswrapper[4787]: I1203 18:31:38.243362 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/b52259b1-98e9-4791-a981-fc62fdd0e138-ceph\") pod \"manila-share-share1-0\" (UID: \"b52259b1-98e9-4791-a981-fc62fdd0e138\") " pod="openstack/manila-share-share1-0" Dec 03 18:31:38 crc kubenswrapper[4787]: I1203 18:31:38.343439 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 03 18:31:38 crc kubenswrapper[4787]: W1203 18:31:38.928538 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb52259b1_98e9_4791_a981_fc62fdd0e138.slice/crio-e496e0695ae53efca5f3b576c8db186aac85d03a8b40034680d6ed7ec7e93e1e WatchSource:0}: Error finding container e496e0695ae53efca5f3b576c8db186aac85d03a8b40034680d6ed7ec7e93e1e: Status 404 returned error can't find the container with id e496e0695ae53efca5f3b576c8db186aac85d03a8b40034680d6ed7ec7e93e1e Dec 03 18:31:38 crc kubenswrapper[4787]: I1203 18:31:38.936738 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 03 18:31:39 crc kubenswrapper[4787]: I1203 18:31:39.018806 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b52259b1-98e9-4791-a981-fc62fdd0e138","Type":"ContainerStarted","Data":"e496e0695ae53efca5f3b576c8db186aac85d03a8b40034680d6ed7ec7e93e1e"} Dec 03 18:31:39 crc kubenswrapper[4787]: I1203 18:31:39.025239 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11460db3-3d3a-426a-9980-f1dd41a84497","Type":"ContainerStarted","Data":"1513ee8e9add0b05a48bc4b9dd085938647520f2b99050a88392a4432bbba436"} Dec 03 18:31:39 crc kubenswrapper[4787]: I1203 18:31:39.025287 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11460db3-3d3a-426a-9980-f1dd41a84497","Type":"ContainerStarted","Data":"261e400a5388fdaff778793755a45d9521a9e15877a3256910e9fb1f499aeb3d"} Dec 03 18:31:39 crc kubenswrapper[4787]: I1203 18:31:39.769684 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:31:39 crc kubenswrapper[4787]: E1203 18:31:39.774379 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:31:40 crc kubenswrapper[4787]: I1203 18:31:40.067391 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b52259b1-98e9-4791-a981-fc62fdd0e138","Type":"ContainerStarted","Data":"1c23bc6be715c3536d29ce08f29fc9e52759842c913ac9568f459d014e7b2012"} Dec 03 18:31:40 crc kubenswrapper[4787]: I1203 18:31:40.070617 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11460db3-3d3a-426a-9980-f1dd41a84497","Type":"ContainerStarted","Data":"3092244206f15c7deb7dfc4d7454b58ef9e19d9684157bc5f30627926570f26c"} Dec 03 18:31:40 crc kubenswrapper[4787]: I1203 18:31:40.917865 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.040227 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-scripts\") pod \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.040351 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-etc-machine-id\") pod \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.040442 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data-custom\") pod \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.040528 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hq722\" (UniqueName: \"kubernetes.io/projected/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-kube-api-access-hq722\") pod \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.040579 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data\") pod \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.040608 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-combined-ca-bundle\") pod \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\" (UID: \"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa\") " Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.041940 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" (UID: "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.048864 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-scripts" (OuterVolumeSpecName: "scripts") pod "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" (UID: "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.049476 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-kube-api-access-hq722" (OuterVolumeSpecName: "kube-api-access-hq722") pod "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" (UID: "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa"). InnerVolumeSpecName "kube-api-access-hq722". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.063441 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" (UID: "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.082730 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"b52259b1-98e9-4791-a981-fc62fdd0e138","Type":"ContainerStarted","Data":"c7d687b7994fd3bba234602c7b01c8dd4d0835e90d2832277cafc122289f6e9c"} Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.089222 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11460db3-3d3a-426a-9980-f1dd41a84497","Type":"ContainerStarted","Data":"fd145b63c841cefd50b222d8f21871122d5ef322d290a65ce097bd99e93cfe3f"} Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.099312 4787 generic.go:334] "Generic (PLEG): container finished" podID="27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" containerID="8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5" exitCode=0 Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.099370 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa","Type":"ContainerDied","Data":"8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5"} Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.099399 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"27dc8abf-cac3-49d5-aa05-08b8d8ed05aa","Type":"ContainerDied","Data":"ef0f84737ff82258933a317fd5b7535c53f607fbd29e5525d8afcddbeaa4ae07"} Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.099415 4787 scope.go:117] "RemoveContainer" containerID="48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.099453 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.109438 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=4.109404701 podStartE2EDuration="4.109404701s" podCreationTimestamp="2025-12-03 18:31:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:41.101284434 +0000 UTC m=+4717.918755393" watchObservedRunningTime="2025-12-03 18:31:41.109404701 +0000 UTC m=+4717.926875660" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.109744 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" (UID: "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.143060 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hq722\" (UniqueName: \"kubernetes.io/projected/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-kube-api-access-hq722\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.143088 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.143098 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.143107 4787 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.143115 4787 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.176292 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data" (OuterVolumeSpecName: "config-data") pod "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" (UID: "27dc8abf-cac3-49d5-aa05-08b8d8ed05aa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.182692 4787 scope.go:117] "RemoveContainer" containerID="8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.204340 4787 scope.go:117] "RemoveContainer" containerID="48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294" Dec 03 18:31:41 crc kubenswrapper[4787]: E1203 18:31:41.205408 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294\": container with ID starting with 48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294 not found: ID does not exist" containerID="48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.205460 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294"} err="failed to get container status \"48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294\": rpc error: code = NotFound desc = could not find container \"48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294\": container with ID starting with 48dd75399d9abc5f8a17005746dd53f325fed493fa682b311c9b3f7184db3294 not found: ID does not exist" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.205483 4787 scope.go:117] "RemoveContainer" containerID="8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5" Dec 03 18:31:41 crc kubenswrapper[4787]: E1203 18:31:41.208362 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5\": container with ID starting with 8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5 not found: ID does not exist" containerID="8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.208392 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5"} err="failed to get container status \"8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5\": rpc error: code = NotFound desc = could not find container \"8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5\": container with ID starting with 8f24293f2e83617f8d224507ef5fa97638ec376e510eb72edf72b966132e8cd5 not found: ID does not exist" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.244919 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.260913 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.441331 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.457913 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.474561 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:31:41 crc kubenswrapper[4787]: E1203 18:31:41.475071 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" containerName="probe" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.475087 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" containerName="probe" Dec 03 18:31:41 crc kubenswrapper[4787]: E1203 18:31:41.475107 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" containerName="manila-scheduler" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.475113 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" containerName="manila-scheduler" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.475300 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" containerName="manila-scheduler" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.475335 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" containerName="probe" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.476474 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.482436 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.490851 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.552548 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pzlv8\" (UniqueName: \"kubernetes.io/projected/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-kube-api-access-pzlv8\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.552629 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.553027 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.553239 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-scripts\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.553424 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.553573 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-config-data\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.655810 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.656287 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-scripts\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.656332 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.656369 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-config-data\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.656444 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pzlv8\" (UniqueName: \"kubernetes.io/projected/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-kube-api-access-pzlv8\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.656466 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.656592 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.660914 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.661297 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-scripts\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.663640 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-config-data\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.664352 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.674276 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pzlv8\" (UniqueName: \"kubernetes.io/projected/5999e180-294a-4f1a-ae8f-e5a7dbf73b0b-kube-api-access-pzlv8\") pod \"manila-scheduler-0\" (UID: \"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b\") " pod="openstack/manila-scheduler-0" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.777598 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27dc8abf-cac3-49d5-aa05-08b8d8ed05aa" path="/var/lib/kubelet/pods/27dc8abf-cac3-49d5-aa05-08b8d8ed05aa/volumes" Dec 03 18:31:41 crc kubenswrapper[4787]: I1203 18:31:41.793431 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 03 18:31:42 crc kubenswrapper[4787]: I1203 18:31:42.115215 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"11460db3-3d3a-426a-9980-f1dd41a84497","Type":"ContainerStarted","Data":"6855b581c477da482e882faeaf70f4daed8bca0c9261eba240ba77f87e2ea71c"} Dec 03 18:31:42 crc kubenswrapper[4787]: I1203 18:31:42.115393 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 03 18:31:42 crc kubenswrapper[4787]: I1203 18:31:42.148005 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.850903824 podStartE2EDuration="5.147982324s" podCreationTimestamp="2025-12-03 18:31:37 +0000 UTC" firstStartedPulling="2025-12-03 18:31:38.223634324 +0000 UTC m=+4715.041105303" lastFinishedPulling="2025-12-03 18:31:41.520712854 +0000 UTC m=+4718.338183803" observedRunningTime="2025-12-03 18:31:42.134281198 +0000 UTC m=+4718.951752167" watchObservedRunningTime="2025-12-03 18:31:42.147982324 +0000 UTC m=+4718.965453283" Dec 03 18:31:42 crc kubenswrapper[4787]: I1203 18:31:42.279662 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 03 18:31:43 crc kubenswrapper[4787]: I1203 18:31:43.131929 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b","Type":"ContainerStarted","Data":"d24abab837e1ec783c12df206cb9d21d22eea4e8f0fe4cd205dca88172031ebe"} Dec 03 18:31:43 crc kubenswrapper[4787]: I1203 18:31:43.132403 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b","Type":"ContainerStarted","Data":"889b4e2b06978dd831940c6925cda629e942cee1c74d6c73774701850341513f"} Dec 03 18:31:44 crc kubenswrapper[4787]: I1203 18:31:44.144327 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"5999e180-294a-4f1a-ae8f-e5a7dbf73b0b","Type":"ContainerStarted","Data":"9954c3cc9e1ac12663210882609a26c7dffbb33e585b991228778ebcef8629af"} Dec 03 18:31:44 crc kubenswrapper[4787]: I1203 18:31:44.163686 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.163664997 podStartE2EDuration="3.163664997s" podCreationTimestamp="2025-12-03 18:31:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:31:44.159830395 +0000 UTC m=+4720.977301354" watchObservedRunningTime="2025-12-03 18:31:44.163664997 +0000 UTC m=+4720.981135956" Dec 03 18:31:45 crc kubenswrapper[4787]: I1203 18:31:45.690653 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-66fb6d975b-t28ks" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon" probeResult="failure" output="Get \"https://10.217.1.71:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.1.71:8443: connect: connection refused" Dec 03 18:31:48 crc kubenswrapper[4787]: I1203 18:31:48.344637 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.231870 4787 generic.go:334] "Generic (PLEG): container finished" podID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerID="422f921f2714b4435cd12b9578e5ee5f5f35831b82e8bae5717d3c307be5667b" exitCode=137 Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.232121 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fb6d975b-t28ks" event={"ID":"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3","Type":"ContainerDied","Data":"422f921f2714b4435cd12b9578e5ee5f5f35831b82e8bae5717d3c307be5667b"} Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.232425 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-66fb6d975b-t28ks" event={"ID":"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3","Type":"ContainerDied","Data":"a30fe7537ceae9aafa3e8b2865cffd07d223f0dcc63e2db857e324409195fa9b"} Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.232449 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a30fe7537ceae9aafa3e8b2865cffd07d223f0dcc63e2db857e324409195fa9b" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.325788 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.431659 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-logs\") pod \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.431934 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-combined-ca-bundle\") pod \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.432146 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-config-data\") pod \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.432886 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-scripts\") pod \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.433007 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-tls-certs\") pod \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.433135 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kw5tk\" (UniqueName: \"kubernetes.io/projected/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-kube-api-access-kw5tk\") pod \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.433248 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-secret-key\") pod \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\" (UID: \"acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3\") " Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.434908 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-logs" (OuterVolumeSpecName: "logs") pod "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" (UID: "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.438876 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" (UID: "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.440583 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-kube-api-access-kw5tk" (OuterVolumeSpecName: "kube-api-access-kw5tk") pod "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" (UID: "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3"). InnerVolumeSpecName "kube-api-access-kw5tk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.463557 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-config-data" (OuterVolumeSpecName: "config-data") pod "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" (UID: "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.468188 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" (UID: "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.483165 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-scripts" (OuterVolumeSpecName: "scripts") pod "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" (UID: "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.522166 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" (UID: "acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.537292 4787 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-logs\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.537327 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.537338 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.537347 4787 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-scripts\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.537369 4787 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.537378 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kw5tk\" (UniqueName: \"kubernetes.io/projected/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-kube-api-access-kw5tk\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.537387 4787 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 03 18:31:51 crc kubenswrapper[4787]: I1203 18:31:51.794566 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 03 18:31:52 crc kubenswrapper[4787]: I1203 18:31:52.241432 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-66fb6d975b-t28ks" Dec 03 18:31:52 crc kubenswrapper[4787]: I1203 18:31:52.270954 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-66fb6d975b-t28ks"] Dec 03 18:31:52 crc kubenswrapper[4787]: I1203 18:31:52.293225 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-66fb6d975b-t28ks"] Dec 03 18:31:52 crc kubenswrapper[4787]: I1203 18:31:52.766461 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:31:52 crc kubenswrapper[4787]: E1203 18:31:52.767003 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:31:53 crc kubenswrapper[4787]: I1203 18:31:53.782177 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" path="/var/lib/kubelet/pods/acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3/volumes" Dec 03 18:31:55 crc kubenswrapper[4787]: I1203 18:31:55.833539 4787 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod7e131b5d-ae30-4381-b9ec-b14feffb63ee"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod7e131b5d-ae30-4381-b9ec-b14feffb63ee] : Timed out while waiting for systemd to remove kubepods-besteffort-pod7e131b5d_ae30_4381_b9ec_b14feffb63ee.slice" Dec 03 18:31:55 crc kubenswrapper[4787]: E1203 18:31:55.833863 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod7e131b5d-ae30-4381-b9ec-b14feffb63ee] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod7e131b5d-ae30-4381-b9ec-b14feffb63ee] : Timed out while waiting for systemd to remove kubepods-besteffort-pod7e131b5d_ae30_4381_b9ec_b14feffb63ee.slice" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" podUID="7e131b5d-ae30-4381-b9ec-b14feffb63ee" Dec 03 18:31:56 crc kubenswrapper[4787]: I1203 18:31:56.285000 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6559847fc9-hfmfg" Dec 03 18:31:56 crc kubenswrapper[4787]: I1203 18:31:56.382823 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-hfmfg"] Dec 03 18:31:56 crc kubenswrapper[4787]: I1203 18:31:56.391957 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6559847fc9-hfmfg"] Dec 03 18:31:57 crc kubenswrapper[4787]: I1203 18:31:57.786568 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7e131b5d-ae30-4381-b9ec-b14feffb63ee" path="/var/lib/kubelet/pods/7e131b5d-ae30-4381-b9ec-b14feffb63ee/volumes" Dec 03 18:31:59 crc kubenswrapper[4787]: I1203 18:31:59.944757 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 03 18:32:03 crc kubenswrapper[4787]: I1203 18:32:03.218976 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 03 18:32:07 crc kubenswrapper[4787]: I1203 18:32:07.427970 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 03 18:32:07 crc kubenswrapper[4787]: I1203 18:32:07.770894 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:32:07 crc kubenswrapper[4787]: E1203 18:32:07.771611 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:32:18 crc kubenswrapper[4787]: I1203 18:32:18.766689 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:32:18 crc kubenswrapper[4787]: E1203 18:32:18.767503 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:32:30 crc kubenswrapper[4787]: I1203 18:32:30.767183 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:32:30 crc kubenswrapper[4787]: E1203 18:32:30.768350 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:32:44 crc kubenswrapper[4787]: I1203 18:32:44.766748 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:32:44 crc kubenswrapper[4787]: E1203 18:32:44.767614 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:32:55 crc kubenswrapper[4787]: I1203 18:32:55.767752 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:32:55 crc kubenswrapper[4787]: E1203 18:32:55.783410 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:33:07 crc kubenswrapper[4787]: I1203 18:33:07.766909 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:33:07 crc kubenswrapper[4787]: E1203 18:33:07.767686 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.211741 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-x244d"] Dec 03 18:33:14 crc kubenswrapper[4787]: E1203 18:33:14.213488 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.213502 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon" Dec 03 18:33:14 crc kubenswrapper[4787]: E1203 18:33:14.213527 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon-log" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.213533 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon-log" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.213746 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon-log" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.213774 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="acabd0b4-b5f4-46a1-b5d1-cfce6f076ff3" containerName="horizon" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.215480 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.239211 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x244d"] Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.309126 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-utilities\") pod \"redhat-operators-x244d\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.309403 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-catalog-content\") pod \"redhat-operators-x244d\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.309557 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gh2z\" (UniqueName: \"kubernetes.io/projected/ac443653-2d20-48bf-aa6d-2641ec616d05-kube-api-access-9gh2z\") pod \"redhat-operators-x244d\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.411706 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-catalog-content\") pod \"redhat-operators-x244d\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.411805 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gh2z\" (UniqueName: \"kubernetes.io/projected/ac443653-2d20-48bf-aa6d-2641ec616d05-kube-api-access-9gh2z\") pod \"redhat-operators-x244d\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.411956 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-utilities\") pod \"redhat-operators-x244d\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.412602 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-utilities\") pod \"redhat-operators-x244d\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.412905 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-catalog-content\") pod \"redhat-operators-x244d\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.433665 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gh2z\" (UniqueName: \"kubernetes.io/projected/ac443653-2d20-48bf-aa6d-2641ec616d05-kube-api-access-9gh2z\") pod \"redhat-operators-x244d\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:14 crc kubenswrapper[4787]: I1203 18:33:14.550092 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:15 crc kubenswrapper[4787]: W1203 18:33:15.068308 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podac443653_2d20_48bf_aa6d_2641ec616d05.slice/crio-25c744eaec47be1f7344af041854da4b4ce5fa3b6407fbff390c671b9606e90a WatchSource:0}: Error finding container 25c744eaec47be1f7344af041854da4b4ce5fa3b6407fbff390c671b9606e90a: Status 404 returned error can't find the container with id 25c744eaec47be1f7344af041854da4b4ce5fa3b6407fbff390c671b9606e90a Dec 03 18:33:15 crc kubenswrapper[4787]: I1203 18:33:15.071582 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-x244d"] Dec 03 18:33:15 crc kubenswrapper[4787]: I1203 18:33:15.251107 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x244d" event={"ID":"ac443653-2d20-48bf-aa6d-2641ec616d05","Type":"ContainerStarted","Data":"25c744eaec47be1f7344af041854da4b4ce5fa3b6407fbff390c671b9606e90a"} Dec 03 18:33:16 crc kubenswrapper[4787]: I1203 18:33:16.405732 4787 generic.go:334] "Generic (PLEG): container finished" podID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerID="e9c218a5b08f7da01efdce3b791b3c4a19adc30045c4175d26ec0e8c677a8dff" exitCode=0 Dec 03 18:33:16 crc kubenswrapper[4787]: I1203 18:33:16.406107 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x244d" event={"ID":"ac443653-2d20-48bf-aa6d-2641ec616d05","Type":"ContainerDied","Data":"e9c218a5b08f7da01efdce3b791b3c4a19adc30045c4175d26ec0e8c677a8dff"} Dec 03 18:33:16 crc kubenswrapper[4787]: I1203 18:33:16.409475 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:33:18 crc kubenswrapper[4787]: I1203 18:33:18.433556 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x244d" event={"ID":"ac443653-2d20-48bf-aa6d-2641ec616d05","Type":"ContainerStarted","Data":"cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb"} Dec 03 18:33:18 crc kubenswrapper[4787]: I1203 18:33:18.766164 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:33:18 crc kubenswrapper[4787]: E1203 18:33:18.766751 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:33:20 crc kubenswrapper[4787]: I1203 18:33:20.463378 4787 generic.go:334] "Generic (PLEG): container finished" podID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerID="cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb" exitCode=0 Dec 03 18:33:20 crc kubenswrapper[4787]: I1203 18:33:20.463487 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x244d" event={"ID":"ac443653-2d20-48bf-aa6d-2641ec616d05","Type":"ContainerDied","Data":"cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb"} Dec 03 18:33:21 crc kubenswrapper[4787]: I1203 18:33:21.481971 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x244d" event={"ID":"ac443653-2d20-48bf-aa6d-2641ec616d05","Type":"ContainerStarted","Data":"4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c"} Dec 03 18:33:21 crc kubenswrapper[4787]: I1203 18:33:21.511185 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-x244d" podStartSLOduration=2.997297863 podStartE2EDuration="7.511160314s" podCreationTimestamp="2025-12-03 18:33:14 +0000 UTC" firstStartedPulling="2025-12-03 18:33:16.40923122 +0000 UTC m=+4813.226702189" lastFinishedPulling="2025-12-03 18:33:20.923093671 +0000 UTC m=+4817.740564640" observedRunningTime="2025-12-03 18:33:21.503911911 +0000 UTC m=+4818.321382880" watchObservedRunningTime="2025-12-03 18:33:21.511160314 +0000 UTC m=+4818.328631313" Dec 03 18:33:24 crc kubenswrapper[4787]: I1203 18:33:24.550577 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:24 crc kubenswrapper[4787]: I1203 18:33:24.551237 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:25 crc kubenswrapper[4787]: I1203 18:33:25.624675 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-x244d" podUID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerName="registry-server" probeResult="failure" output=< Dec 03 18:33:25 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 18:33:25 crc kubenswrapper[4787]: > Dec 03 18:33:32 crc kubenswrapper[4787]: I1203 18:33:32.765798 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:33:32 crc kubenswrapper[4787]: E1203 18:33:32.766540 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:33:34 crc kubenswrapper[4787]: I1203 18:33:34.618924 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:34 crc kubenswrapper[4787]: I1203 18:33:34.678557 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:34 crc kubenswrapper[4787]: I1203 18:33:34.863879 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x244d"] Dec 03 18:33:35 crc kubenswrapper[4787]: I1203 18:33:35.658040 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-x244d" podUID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerName="registry-server" containerID="cri-o://4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c" gracePeriod=2 Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.182500 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.291795 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-utilities\") pod \"ac443653-2d20-48bf-aa6d-2641ec616d05\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.291956 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9gh2z\" (UniqueName: \"kubernetes.io/projected/ac443653-2d20-48bf-aa6d-2641ec616d05-kube-api-access-9gh2z\") pod \"ac443653-2d20-48bf-aa6d-2641ec616d05\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.292095 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-catalog-content\") pod \"ac443653-2d20-48bf-aa6d-2641ec616d05\" (UID: \"ac443653-2d20-48bf-aa6d-2641ec616d05\") " Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.293067 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-utilities" (OuterVolumeSpecName: "utilities") pod "ac443653-2d20-48bf-aa6d-2641ec616d05" (UID: "ac443653-2d20-48bf-aa6d-2641ec616d05"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.300544 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac443653-2d20-48bf-aa6d-2641ec616d05-kube-api-access-9gh2z" (OuterVolumeSpecName: "kube-api-access-9gh2z") pod "ac443653-2d20-48bf-aa6d-2641ec616d05" (UID: "ac443653-2d20-48bf-aa6d-2641ec616d05"). InnerVolumeSpecName "kube-api-access-9gh2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.395923 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.396160 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9gh2z\" (UniqueName: \"kubernetes.io/projected/ac443653-2d20-48bf-aa6d-2641ec616d05-kube-api-access-9gh2z\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.431862 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ac443653-2d20-48bf-aa6d-2641ec616d05" (UID: "ac443653-2d20-48bf-aa6d-2641ec616d05"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.498115 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ac443653-2d20-48bf-aa6d-2641ec616d05-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.671276 4787 generic.go:334] "Generic (PLEG): container finished" podID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerID="4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c" exitCode=0 Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.671321 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x244d" event={"ID":"ac443653-2d20-48bf-aa6d-2641ec616d05","Type":"ContainerDied","Data":"4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c"} Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.671348 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-x244d" event={"ID":"ac443653-2d20-48bf-aa6d-2641ec616d05","Type":"ContainerDied","Data":"25c744eaec47be1f7344af041854da4b4ce5fa3b6407fbff390c671b9606e90a"} Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.671366 4787 scope.go:117] "RemoveContainer" containerID="4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.671492 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-x244d" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.700185 4787 scope.go:117] "RemoveContainer" containerID="cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.706732 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-x244d"] Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.717197 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-x244d"] Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.758963 4787 scope.go:117] "RemoveContainer" containerID="e9c218a5b08f7da01efdce3b791b3c4a19adc30045c4175d26ec0e8c677a8dff" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.785570 4787 scope.go:117] "RemoveContainer" containerID="4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c" Dec 03 18:33:36 crc kubenswrapper[4787]: E1203 18:33:36.788256 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c\": container with ID starting with 4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c not found: ID does not exist" containerID="4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.788291 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c"} err="failed to get container status \"4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c\": rpc error: code = NotFound desc = could not find container \"4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c\": container with ID starting with 4dcb1b51129d174cdd93202f68ee5b2a5004b8362f5577c9c719b0b7f581267c not found: ID does not exist" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.788313 4787 scope.go:117] "RemoveContainer" containerID="cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb" Dec 03 18:33:36 crc kubenswrapper[4787]: E1203 18:33:36.793620 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb\": container with ID starting with cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb not found: ID does not exist" containerID="cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.793724 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb"} err="failed to get container status \"cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb\": rpc error: code = NotFound desc = could not find container \"cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb\": container with ID starting with cc37d5c9cfabe5de0c19cf6f4346c63f682b2e01ca5296f8f09ae75d1ffa9ffb not found: ID does not exist" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.793751 4787 scope.go:117] "RemoveContainer" containerID="e9c218a5b08f7da01efdce3b791b3c4a19adc30045c4175d26ec0e8c677a8dff" Dec 03 18:33:36 crc kubenswrapper[4787]: E1203 18:33:36.794936 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9c218a5b08f7da01efdce3b791b3c4a19adc30045c4175d26ec0e8c677a8dff\": container with ID starting with e9c218a5b08f7da01efdce3b791b3c4a19adc30045c4175d26ec0e8c677a8dff not found: ID does not exist" containerID="e9c218a5b08f7da01efdce3b791b3c4a19adc30045c4175d26ec0e8c677a8dff" Dec 03 18:33:36 crc kubenswrapper[4787]: I1203 18:33:36.794975 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9c218a5b08f7da01efdce3b791b3c4a19adc30045c4175d26ec0e8c677a8dff"} err="failed to get container status \"e9c218a5b08f7da01efdce3b791b3c4a19adc30045c4175d26ec0e8c677a8dff\": rpc error: code = NotFound desc = could not find container \"e9c218a5b08f7da01efdce3b791b3c4a19adc30045c4175d26ec0e8c677a8dff\": container with ID starting with e9c218a5b08f7da01efdce3b791b3c4a19adc30045c4175d26ec0e8c677a8dff not found: ID does not exist" Dec 03 18:33:37 crc kubenswrapper[4787]: I1203 18:33:37.815993 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ac443653-2d20-48bf-aa6d-2641ec616d05" path="/var/lib/kubelet/pods/ac443653-2d20-48bf-aa6d-2641ec616d05/volumes" Dec 03 18:33:44 crc kubenswrapper[4787]: I1203 18:33:44.766726 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:33:44 crc kubenswrapper[4787]: E1203 18:33:44.767499 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:33:57 crc kubenswrapper[4787]: I1203 18:33:57.766708 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:33:57 crc kubenswrapper[4787]: E1203 18:33:57.767887 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.684886 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cvklc"] Dec 03 18:34:01 crc kubenswrapper[4787]: E1203 18:34:01.686359 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerName="registry-server" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.686394 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerName="registry-server" Dec 03 18:34:01 crc kubenswrapper[4787]: E1203 18:34:01.686472 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerName="extract-content" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.686488 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerName="extract-content" Dec 03 18:34:01 crc kubenswrapper[4787]: E1203 18:34:01.686504 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerName="extract-utilities" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.686518 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerName="extract-utilities" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.686981 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac443653-2d20-48bf-aa6d-2641ec616d05" containerName="registry-server" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.690641 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.705885 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cvklc"] Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.862039 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-utilities\") pod \"community-operators-cvklc\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.862134 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-catalog-content\") pod \"community-operators-cvklc\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.863115 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pwcv\" (UniqueName: \"kubernetes.io/projected/3557b99b-6437-4b2d-b633-7969430ff38e-kube-api-access-8pwcv\") pod \"community-operators-cvklc\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.965064 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-utilities\") pod \"community-operators-cvklc\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.965126 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-catalog-content\") pod \"community-operators-cvklc\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.965216 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pwcv\" (UniqueName: \"kubernetes.io/projected/3557b99b-6437-4b2d-b633-7969430ff38e-kube-api-access-8pwcv\") pod \"community-operators-cvklc\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.966067 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-utilities\") pod \"community-operators-cvklc\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.966104 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-catalog-content\") pod \"community-operators-cvklc\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:01 crc kubenswrapper[4787]: I1203 18:34:01.999551 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pwcv\" (UniqueName: \"kubernetes.io/projected/3557b99b-6437-4b2d-b633-7969430ff38e-kube-api-access-8pwcv\") pod \"community-operators-cvklc\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:02 crc kubenswrapper[4787]: I1203 18:34:02.031597 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:02 crc kubenswrapper[4787]: I1203 18:34:02.656495 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cvklc"] Dec 03 18:34:03 crc kubenswrapper[4787]: I1203 18:34:03.094447 4787 generic.go:334] "Generic (PLEG): container finished" podID="3557b99b-6437-4b2d-b633-7969430ff38e" containerID="8711e2ef5b6eabeed5be9a12e66529276667806aef4fc081d625eb65872b1cf3" exitCode=0 Dec 03 18:34:03 crc kubenswrapper[4787]: I1203 18:34:03.094850 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvklc" event={"ID":"3557b99b-6437-4b2d-b633-7969430ff38e","Type":"ContainerDied","Data":"8711e2ef5b6eabeed5be9a12e66529276667806aef4fc081d625eb65872b1cf3"} Dec 03 18:34:03 crc kubenswrapper[4787]: I1203 18:34:03.094891 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvklc" event={"ID":"3557b99b-6437-4b2d-b633-7969430ff38e","Type":"ContainerStarted","Data":"46dbb99abb868aa0bd782284410caffb52480a183b36bc63c796a5eb669c3987"} Dec 03 18:34:04 crc kubenswrapper[4787]: I1203 18:34:04.110337 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvklc" event={"ID":"3557b99b-6437-4b2d-b633-7969430ff38e","Type":"ContainerStarted","Data":"9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75"} Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.121834 4787 generic.go:334] "Generic (PLEG): container finished" podID="3557b99b-6437-4b2d-b633-7969430ff38e" containerID="9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75" exitCode=0 Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.121900 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvklc" event={"ID":"3557b99b-6437-4b2d-b633-7969430ff38e","Type":"ContainerDied","Data":"9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75"} Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.676480 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lzztr"] Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.683598 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.698617 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzztr"] Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.860484 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-catalog-content\") pod \"redhat-marketplace-lzztr\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.860842 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9l58r\" (UniqueName: \"kubernetes.io/projected/b284e821-07e0-4f3a-aa78-2a67a637f1df-kube-api-access-9l58r\") pod \"redhat-marketplace-lzztr\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.860881 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-utilities\") pod \"redhat-marketplace-lzztr\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.962730 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-catalog-content\") pod \"redhat-marketplace-lzztr\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.963270 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-catalog-content\") pod \"redhat-marketplace-lzztr\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.963659 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9l58r\" (UniqueName: \"kubernetes.io/projected/b284e821-07e0-4f3a-aa78-2a67a637f1df-kube-api-access-9l58r\") pod \"redhat-marketplace-lzztr\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.963776 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-utilities\") pod \"redhat-marketplace-lzztr\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:05 crc kubenswrapper[4787]: I1203 18:34:05.964204 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-utilities\") pod \"redhat-marketplace-lzztr\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:06 crc kubenswrapper[4787]: I1203 18:34:06.134189 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvklc" event={"ID":"3557b99b-6437-4b2d-b633-7969430ff38e","Type":"ContainerStarted","Data":"be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902"} Dec 03 18:34:06 crc kubenswrapper[4787]: I1203 18:34:06.158608 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cvklc" podStartSLOduration=2.738144787 podStartE2EDuration="5.158591609s" podCreationTimestamp="2025-12-03 18:34:01 +0000 UTC" firstStartedPulling="2025-12-03 18:34:03.103665896 +0000 UTC m=+4859.921136855" lastFinishedPulling="2025-12-03 18:34:05.524112688 +0000 UTC m=+4862.341583677" observedRunningTime="2025-12-03 18:34:06.151249923 +0000 UTC m=+4862.968720892" watchObservedRunningTime="2025-12-03 18:34:06.158591609 +0000 UTC m=+4862.976062568" Dec 03 18:34:06 crc kubenswrapper[4787]: I1203 18:34:06.612742 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9l58r\" (UniqueName: \"kubernetes.io/projected/b284e821-07e0-4f3a-aa78-2a67a637f1df-kube-api-access-9l58r\") pod \"redhat-marketplace-lzztr\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:06 crc kubenswrapper[4787]: I1203 18:34:06.633228 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:07 crc kubenswrapper[4787]: I1203 18:34:07.129901 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzztr"] Dec 03 18:34:07 crc kubenswrapper[4787]: I1203 18:34:07.148762 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzztr" event={"ID":"b284e821-07e0-4f3a-aa78-2a67a637f1df","Type":"ContainerStarted","Data":"69cca219a7aaaf561ebc0f15e3cacfff460ec53376fc5fba1be5cf4d28ec173c"} Dec 03 18:34:08 crc kubenswrapper[4787]: I1203 18:34:08.161445 4787 generic.go:334] "Generic (PLEG): container finished" podID="b284e821-07e0-4f3a-aa78-2a67a637f1df" containerID="7bfad043263ae5cabad92daaf7767fd2c78bab2b3f21bd34b139d3a3396421fb" exitCode=0 Dec 03 18:34:08 crc kubenswrapper[4787]: I1203 18:34:08.161530 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzztr" event={"ID":"b284e821-07e0-4f3a-aa78-2a67a637f1df","Type":"ContainerDied","Data":"7bfad043263ae5cabad92daaf7767fd2c78bab2b3f21bd34b139d3a3396421fb"} Dec 03 18:34:08 crc kubenswrapper[4787]: I1203 18:34:08.766711 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:34:08 crc kubenswrapper[4787]: E1203 18:34:08.767381 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:34:10 crc kubenswrapper[4787]: I1203 18:34:10.188291 4787 generic.go:334] "Generic (PLEG): container finished" podID="b284e821-07e0-4f3a-aa78-2a67a637f1df" containerID="df92c7fa4b723849060aa471c3cb816788093ff85d19b730d6c77126eb86aeb9" exitCode=0 Dec 03 18:34:10 crc kubenswrapper[4787]: I1203 18:34:10.188445 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzztr" event={"ID":"b284e821-07e0-4f3a-aa78-2a67a637f1df","Type":"ContainerDied","Data":"df92c7fa4b723849060aa471c3cb816788093ff85d19b730d6c77126eb86aeb9"} Dec 03 18:34:11 crc kubenswrapper[4787]: I1203 18:34:11.200946 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzztr" event={"ID":"b284e821-07e0-4f3a-aa78-2a67a637f1df","Type":"ContainerStarted","Data":"308dca2edb8fa316617f469a4ee264d640968ed1d5e4ebc035b0cba6dcd2e4fc"} Dec 03 18:34:11 crc kubenswrapper[4787]: I1203 18:34:11.240608 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lzztr" podStartSLOduration=3.759411198 podStartE2EDuration="6.240582671s" podCreationTimestamp="2025-12-03 18:34:05 +0000 UTC" firstStartedPulling="2025-12-03 18:34:08.165120539 +0000 UTC m=+4864.982591498" lastFinishedPulling="2025-12-03 18:34:10.646291972 +0000 UTC m=+4867.463762971" observedRunningTime="2025-12-03 18:34:11.223784983 +0000 UTC m=+4868.041255942" watchObservedRunningTime="2025-12-03 18:34:11.240582671 +0000 UTC m=+4868.058053630" Dec 03 18:34:12 crc kubenswrapper[4787]: I1203 18:34:12.032178 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:12 crc kubenswrapper[4787]: I1203 18:34:12.032588 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:12 crc kubenswrapper[4787]: I1203 18:34:12.092999 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:12 crc kubenswrapper[4787]: I1203 18:34:12.261904 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:14 crc kubenswrapper[4787]: I1203 18:34:14.458588 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cvklc"] Dec 03 18:34:14 crc kubenswrapper[4787]: I1203 18:34:14.459119 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cvklc" podUID="3557b99b-6437-4b2d-b633-7969430ff38e" containerName="registry-server" containerID="cri-o://be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902" gracePeriod=2 Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.237698 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.265848 4787 generic.go:334] "Generic (PLEG): container finished" podID="3557b99b-6437-4b2d-b633-7969430ff38e" containerID="be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902" exitCode=0 Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.265893 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvklc" event={"ID":"3557b99b-6437-4b2d-b633-7969430ff38e","Type":"ContainerDied","Data":"be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902"} Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.265920 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cvklc" event={"ID":"3557b99b-6437-4b2d-b633-7969430ff38e","Type":"ContainerDied","Data":"46dbb99abb868aa0bd782284410caffb52480a183b36bc63c796a5eb669c3987"} Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.265921 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cvklc" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.265938 4787 scope.go:117] "RemoveContainer" containerID="be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.295864 4787 scope.go:117] "RemoveContainer" containerID="9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.328398 4787 scope.go:117] "RemoveContainer" containerID="8711e2ef5b6eabeed5be9a12e66529276667806aef4fc081d625eb65872b1cf3" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.334999 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-utilities\") pod \"3557b99b-6437-4b2d-b633-7969430ff38e\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.335111 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8pwcv\" (UniqueName: \"kubernetes.io/projected/3557b99b-6437-4b2d-b633-7969430ff38e-kube-api-access-8pwcv\") pod \"3557b99b-6437-4b2d-b633-7969430ff38e\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.335197 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-catalog-content\") pod \"3557b99b-6437-4b2d-b633-7969430ff38e\" (UID: \"3557b99b-6437-4b2d-b633-7969430ff38e\") " Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.339275 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-utilities" (OuterVolumeSpecName: "utilities") pod "3557b99b-6437-4b2d-b633-7969430ff38e" (UID: "3557b99b-6437-4b2d-b633-7969430ff38e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.348404 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3557b99b-6437-4b2d-b633-7969430ff38e-kube-api-access-8pwcv" (OuterVolumeSpecName: "kube-api-access-8pwcv") pod "3557b99b-6437-4b2d-b633-7969430ff38e" (UID: "3557b99b-6437-4b2d-b633-7969430ff38e"). InnerVolumeSpecName "kube-api-access-8pwcv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.385903 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3557b99b-6437-4b2d-b633-7969430ff38e" (UID: "3557b99b-6437-4b2d-b633-7969430ff38e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.416625 4787 scope.go:117] "RemoveContainer" containerID="be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902" Dec 03 18:34:16 crc kubenswrapper[4787]: E1203 18:34:16.417133 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902\": container with ID starting with be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902 not found: ID does not exist" containerID="be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.417185 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902"} err="failed to get container status \"be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902\": rpc error: code = NotFound desc = could not find container \"be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902\": container with ID starting with be50ab0eace0d6f6fb7736132157f3d034a080210f905ca37421134c8cc18902 not found: ID does not exist" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.417220 4787 scope.go:117] "RemoveContainer" containerID="9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75" Dec 03 18:34:16 crc kubenswrapper[4787]: E1203 18:34:16.417553 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75\": container with ID starting with 9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75 not found: ID does not exist" containerID="9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.417603 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75"} err="failed to get container status \"9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75\": rpc error: code = NotFound desc = could not find container \"9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75\": container with ID starting with 9f1d332801c048433325afaa69ee44f34da29541c7de5e80430ae8e0c78c9a75 not found: ID does not exist" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.417630 4787 scope.go:117] "RemoveContainer" containerID="8711e2ef5b6eabeed5be9a12e66529276667806aef4fc081d625eb65872b1cf3" Dec 03 18:34:16 crc kubenswrapper[4787]: E1203 18:34:16.418191 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8711e2ef5b6eabeed5be9a12e66529276667806aef4fc081d625eb65872b1cf3\": container with ID starting with 8711e2ef5b6eabeed5be9a12e66529276667806aef4fc081d625eb65872b1cf3 not found: ID does not exist" containerID="8711e2ef5b6eabeed5be9a12e66529276667806aef4fc081d625eb65872b1cf3" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.418225 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8711e2ef5b6eabeed5be9a12e66529276667806aef4fc081d625eb65872b1cf3"} err="failed to get container status \"8711e2ef5b6eabeed5be9a12e66529276667806aef4fc081d625eb65872b1cf3\": rpc error: code = NotFound desc = could not find container \"8711e2ef5b6eabeed5be9a12e66529276667806aef4fc081d625eb65872b1cf3\": container with ID starting with 8711e2ef5b6eabeed5be9a12e66529276667806aef4fc081d625eb65872b1cf3 not found: ID does not exist" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.438043 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8pwcv\" (UniqueName: \"kubernetes.io/projected/3557b99b-6437-4b2d-b633-7969430ff38e-kube-api-access-8pwcv\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.438093 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.438106 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3557b99b-6437-4b2d-b633-7969430ff38e-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.620217 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cvklc"] Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.634224 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.634289 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.634971 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cvklc"] Dec 03 18:34:16 crc kubenswrapper[4787]: I1203 18:34:16.700994 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:17 crc kubenswrapper[4787]: I1203 18:34:17.346702 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:17 crc kubenswrapper[4787]: I1203 18:34:17.789105 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3557b99b-6437-4b2d-b633-7969430ff38e" path="/var/lib/kubelet/pods/3557b99b-6437-4b2d-b633-7969430ff38e/volumes" Dec 03 18:34:19 crc kubenswrapper[4787]: I1203 18:34:19.456158 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzztr"] Dec 03 18:34:19 crc kubenswrapper[4787]: I1203 18:34:19.456669 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lzztr" podUID="b284e821-07e0-4f3a-aa78-2a67a637f1df" containerName="registry-server" containerID="cri-o://308dca2edb8fa316617f469a4ee264d640968ed1d5e4ebc035b0cba6dcd2e4fc" gracePeriod=2 Dec 03 18:34:20 crc kubenswrapper[4787]: I1203 18:34:20.324277 4787 generic.go:334] "Generic (PLEG): container finished" podID="b284e821-07e0-4f3a-aa78-2a67a637f1df" containerID="308dca2edb8fa316617f469a4ee264d640968ed1d5e4ebc035b0cba6dcd2e4fc" exitCode=0 Dec 03 18:34:20 crc kubenswrapper[4787]: I1203 18:34:20.324371 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzztr" event={"ID":"b284e821-07e0-4f3a-aa78-2a67a637f1df","Type":"ContainerDied","Data":"308dca2edb8fa316617f469a4ee264d640968ed1d5e4ebc035b0cba6dcd2e4fc"} Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.126339 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.252725 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-catalog-content\") pod \"b284e821-07e0-4f3a-aa78-2a67a637f1df\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.252798 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9l58r\" (UniqueName: \"kubernetes.io/projected/b284e821-07e0-4f3a-aa78-2a67a637f1df-kube-api-access-9l58r\") pod \"b284e821-07e0-4f3a-aa78-2a67a637f1df\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.252958 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-utilities\") pod \"b284e821-07e0-4f3a-aa78-2a67a637f1df\" (UID: \"b284e821-07e0-4f3a-aa78-2a67a637f1df\") " Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.253712 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-utilities" (OuterVolumeSpecName: "utilities") pod "b284e821-07e0-4f3a-aa78-2a67a637f1df" (UID: "b284e821-07e0-4f3a-aa78-2a67a637f1df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.260278 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b284e821-07e0-4f3a-aa78-2a67a637f1df-kube-api-access-9l58r" (OuterVolumeSpecName: "kube-api-access-9l58r") pod "b284e821-07e0-4f3a-aa78-2a67a637f1df" (UID: "b284e821-07e0-4f3a-aa78-2a67a637f1df"). InnerVolumeSpecName "kube-api-access-9l58r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.274507 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b284e821-07e0-4f3a-aa78-2a67a637f1df" (UID: "b284e821-07e0-4f3a-aa78-2a67a637f1df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.338039 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lzztr" event={"ID":"b284e821-07e0-4f3a-aa78-2a67a637f1df","Type":"ContainerDied","Data":"69cca219a7aaaf561ebc0f15e3cacfff460ec53376fc5fba1be5cf4d28ec173c"} Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.338110 4787 scope.go:117] "RemoveContainer" containerID="308dca2edb8fa316617f469a4ee264d640968ed1d5e4ebc035b0cba6dcd2e4fc" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.338115 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lzztr" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.355453 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.355717 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9l58r\" (UniqueName: \"kubernetes.io/projected/b284e821-07e0-4f3a-aa78-2a67a637f1df-kube-api-access-9l58r\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.355731 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b284e821-07e0-4f3a-aa78-2a67a637f1df-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.361727 4787 scope.go:117] "RemoveContainer" containerID="df92c7fa4b723849060aa471c3cb816788093ff85d19b730d6c77126eb86aeb9" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.396236 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzztr"] Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.406508 4787 scope.go:117] "RemoveContainer" containerID="7bfad043263ae5cabad92daaf7767fd2c78bab2b3f21bd34b139d3a3396421fb" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.407802 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lzztr"] Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.766415 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:34:21 crc kubenswrapper[4787]: I1203 18:34:21.783538 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b284e821-07e0-4f3a-aa78-2a67a637f1df" path="/var/lib/kubelet/pods/b284e821-07e0-4f3a-aa78-2a67a637f1df/volumes" Dec 03 18:34:22 crc kubenswrapper[4787]: I1203 18:34:22.352503 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"743ebb9764dc004d652685c104dc6240ac82ed6fb6fb7bd45471642d78f70ab2"} Dec 03 18:36:48 crc kubenswrapper[4787]: I1203 18:36:48.989927 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:36:48 crc kubenswrapper[4787]: I1203 18:36:48.990592 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:37:18 crc kubenswrapper[4787]: I1203 18:37:18.989762 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:37:18 crc kubenswrapper[4787]: I1203 18:37:18.990422 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:37:27 crc kubenswrapper[4787]: I1203 18:37:27.801658 4787 scope.go:117] "RemoveContainer" containerID="44090e1deec9d67815c478077aa3038893e46eb258ed036a6788c45da9d41c5b" Dec 03 18:37:27 crc kubenswrapper[4787]: I1203 18:37:27.867857 4787 scope.go:117] "RemoveContainer" containerID="422f921f2714b4435cd12b9578e5ee5f5f35831b82e8bae5717d3c307be5667b" Dec 03 18:37:27 crc kubenswrapper[4787]: I1203 18:37:27.893307 4787 scope.go:117] "RemoveContainer" containerID="2e213c081473b6f716a628245ed15b68bbc0a9d5d80322447926689d2b594b05" Dec 03 18:37:27 crc kubenswrapper[4787]: I1203 18:37:27.925899 4787 scope.go:117] "RemoveContainer" containerID="4de5b41a3df3c3e88ba31d7984a50c7e78af565c2be91b7f608830527d0c964a" Dec 03 18:37:48 crc kubenswrapper[4787]: I1203 18:37:48.990363 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:37:48 crc kubenswrapper[4787]: I1203 18:37:48.991163 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:37:48 crc kubenswrapper[4787]: I1203 18:37:48.991238 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 18:37:48 crc kubenswrapper[4787]: I1203 18:37:48.992468 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"743ebb9764dc004d652685c104dc6240ac82ed6fb6fb7bd45471642d78f70ab2"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:37:48 crc kubenswrapper[4787]: I1203 18:37:48.992577 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://743ebb9764dc004d652685c104dc6240ac82ed6fb6fb7bd45471642d78f70ab2" gracePeriod=600 Dec 03 18:37:49 crc kubenswrapper[4787]: I1203 18:37:49.986200 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="743ebb9764dc004d652685c104dc6240ac82ed6fb6fb7bd45471642d78f70ab2" exitCode=0 Dec 03 18:37:49 crc kubenswrapper[4787]: I1203 18:37:49.986558 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"743ebb9764dc004d652685c104dc6240ac82ed6fb6fb7bd45471642d78f70ab2"} Dec 03 18:37:49 crc kubenswrapper[4787]: I1203 18:37:49.986597 4787 scope.go:117] "RemoveContainer" containerID="2ddd603cfdf0c2d819576d95111529d05fd13b6658fc9e66daca5e25879e78f3" Dec 03 18:37:51 crc kubenswrapper[4787]: I1203 18:37:51.007451 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291"} Dec 03 18:38:28 crc kubenswrapper[4787]: I1203 18:38:28.187962 4787 scope.go:117] "RemoveContainer" containerID="db9f2bebed810a317367c02a17c64f640ced191f4a650a3b4061021e13ea4f8e" Dec 03 18:38:28 crc kubenswrapper[4787]: I1203 18:38:28.222976 4787 scope.go:117] "RemoveContainer" containerID="ddcb4b3a16d83bcc83aa9e5774cd76a4d8a1a7ce755a77d866dd41d9ff9b5e20" Dec 03 18:38:54 crc kubenswrapper[4787]: E1203 18:38:54.380469 4787 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.65:38378->38.102.83.65:44689: write tcp 38.102.83.65:38378->38.102.83.65:44689: write: connection reset by peer Dec 03 18:40:18 crc kubenswrapper[4787]: I1203 18:40:18.989892 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:40:18 crc kubenswrapper[4787]: I1203 18:40:18.990449 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.500777 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-s5lm4"] Dec 03 18:40:22 crc kubenswrapper[4787]: E1203 18:40:22.501788 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3557b99b-6437-4b2d-b633-7969430ff38e" containerName="extract-utilities" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.501803 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3557b99b-6437-4b2d-b633-7969430ff38e" containerName="extract-utilities" Dec 03 18:40:22 crc kubenswrapper[4787]: E1203 18:40:22.501815 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3557b99b-6437-4b2d-b633-7969430ff38e" containerName="extract-content" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.501825 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3557b99b-6437-4b2d-b633-7969430ff38e" containerName="extract-content" Dec 03 18:40:22 crc kubenswrapper[4787]: E1203 18:40:22.501849 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b284e821-07e0-4f3a-aa78-2a67a637f1df" containerName="extract-content" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.501857 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b284e821-07e0-4f3a-aa78-2a67a637f1df" containerName="extract-content" Dec 03 18:40:22 crc kubenswrapper[4787]: E1203 18:40:22.501890 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b284e821-07e0-4f3a-aa78-2a67a637f1df" containerName="extract-utilities" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.501900 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b284e821-07e0-4f3a-aa78-2a67a637f1df" containerName="extract-utilities" Dec 03 18:40:22 crc kubenswrapper[4787]: E1203 18:40:22.501912 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b284e821-07e0-4f3a-aa78-2a67a637f1df" containerName="registry-server" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.501919 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b284e821-07e0-4f3a-aa78-2a67a637f1df" containerName="registry-server" Dec 03 18:40:22 crc kubenswrapper[4787]: E1203 18:40:22.501952 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3557b99b-6437-4b2d-b633-7969430ff38e" containerName="registry-server" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.501959 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="3557b99b-6437-4b2d-b633-7969430ff38e" containerName="registry-server" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.502232 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b284e821-07e0-4f3a-aa78-2a67a637f1df" containerName="registry-server" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.502267 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="3557b99b-6437-4b2d-b633-7969430ff38e" containerName="registry-server" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.504070 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.564313 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s5lm4"] Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.599527 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-utilities\") pod \"certified-operators-s5lm4\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.599766 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4zf6g\" (UniqueName: \"kubernetes.io/projected/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-kube-api-access-4zf6g\") pod \"certified-operators-s5lm4\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.599830 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-catalog-content\") pod \"certified-operators-s5lm4\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.704358 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-utilities\") pod \"certified-operators-s5lm4\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.704535 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4zf6g\" (UniqueName: \"kubernetes.io/projected/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-kube-api-access-4zf6g\") pod \"certified-operators-s5lm4\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.704576 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-catalog-content\") pod \"certified-operators-s5lm4\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.705055 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-catalog-content\") pod \"certified-operators-s5lm4\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.705274 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-utilities\") pod \"certified-operators-s5lm4\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.733795 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4zf6g\" (UniqueName: \"kubernetes.io/projected/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-kube-api-access-4zf6g\") pod \"certified-operators-s5lm4\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:22 crc kubenswrapper[4787]: I1203 18:40:22.831651 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:23 crc kubenswrapper[4787]: I1203 18:40:23.454382 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-s5lm4"] Dec 03 18:40:24 crc kubenswrapper[4787]: I1203 18:40:24.299934 4787 generic.go:334] "Generic (PLEG): container finished" podID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" containerID="fd08cb45146eab40252b2452a42e5aa9541c56dd4132484051f58a0cc9626b0c" exitCode=0 Dec 03 18:40:24 crc kubenswrapper[4787]: I1203 18:40:24.300059 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5lm4" event={"ID":"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff","Type":"ContainerDied","Data":"fd08cb45146eab40252b2452a42e5aa9541c56dd4132484051f58a0cc9626b0c"} Dec 03 18:40:24 crc kubenswrapper[4787]: I1203 18:40:24.300450 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5lm4" event={"ID":"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff","Type":"ContainerStarted","Data":"134920907f37e61e6a4ac5d0c240e7955e9ba394f1733e9552addacfdccb7ce9"} Dec 03 18:40:24 crc kubenswrapper[4787]: I1203 18:40:24.303333 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:40:25 crc kubenswrapper[4787]: I1203 18:40:25.313658 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5lm4" event={"ID":"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff","Type":"ContainerStarted","Data":"25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727"} Dec 03 18:40:26 crc kubenswrapper[4787]: I1203 18:40:26.326455 4787 generic.go:334] "Generic (PLEG): container finished" podID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" containerID="25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727" exitCode=0 Dec 03 18:40:26 crc kubenswrapper[4787]: I1203 18:40:26.326582 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5lm4" event={"ID":"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff","Type":"ContainerDied","Data":"25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727"} Dec 03 18:40:27 crc kubenswrapper[4787]: I1203 18:40:27.339227 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5lm4" event={"ID":"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff","Type":"ContainerStarted","Data":"2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28"} Dec 03 18:40:27 crc kubenswrapper[4787]: I1203 18:40:27.365170 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-s5lm4" podStartSLOduration=2.773769281 podStartE2EDuration="5.365147464s" podCreationTimestamp="2025-12-03 18:40:22 +0000 UTC" firstStartedPulling="2025-12-03 18:40:24.303072221 +0000 UTC m=+5241.120543190" lastFinishedPulling="2025-12-03 18:40:26.894450374 +0000 UTC m=+5243.711921373" observedRunningTime="2025-12-03 18:40:27.356075162 +0000 UTC m=+5244.173546121" watchObservedRunningTime="2025-12-03 18:40:27.365147464 +0000 UTC m=+5244.182618443" Dec 03 18:40:32 crc kubenswrapper[4787]: I1203 18:40:32.832694 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:32 crc kubenswrapper[4787]: I1203 18:40:32.833372 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:32 crc kubenswrapper[4787]: I1203 18:40:32.925147 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:33 crc kubenswrapper[4787]: I1203 18:40:33.975040 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:34 crc kubenswrapper[4787]: I1203 18:40:34.033501 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s5lm4"] Dec 03 18:40:35 crc kubenswrapper[4787]: I1203 18:40:35.433591 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-s5lm4" podUID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" containerName="registry-server" containerID="cri-o://2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28" gracePeriod=2 Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.074999 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.173488 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4zf6g\" (UniqueName: \"kubernetes.io/projected/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-kube-api-access-4zf6g\") pod \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.173642 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-utilities\") pod \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.173748 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-catalog-content\") pod \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\" (UID: \"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff\") " Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.175138 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-utilities" (OuterVolumeSpecName: "utilities") pod "afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" (UID: "afa83387-b414-4fcf-9a8a-1ec08ef8d6ff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.181700 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-kube-api-access-4zf6g" (OuterVolumeSpecName: "kube-api-access-4zf6g") pod "afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" (UID: "afa83387-b414-4fcf-9a8a-1ec08ef8d6ff"). InnerVolumeSpecName "kube-api-access-4zf6g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.269841 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" (UID: "afa83387-b414-4fcf-9a8a-1ec08ef8d6ff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.277481 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4zf6g\" (UniqueName: \"kubernetes.io/projected/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-kube-api-access-4zf6g\") on node \"crc\" DevicePath \"\"" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.277526 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.277556 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.445922 4787 generic.go:334] "Generic (PLEG): container finished" podID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" containerID="2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28" exitCode=0 Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.445963 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5lm4" event={"ID":"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff","Type":"ContainerDied","Data":"2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28"} Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.445992 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-s5lm4" event={"ID":"afa83387-b414-4fcf-9a8a-1ec08ef8d6ff","Type":"ContainerDied","Data":"134920907f37e61e6a4ac5d0c240e7955e9ba394f1733e9552addacfdccb7ce9"} Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.446034 4787 scope.go:117] "RemoveContainer" containerID="2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.446002 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-s5lm4" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.478939 4787 scope.go:117] "RemoveContainer" containerID="25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.485286 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-s5lm4"] Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.512315 4787 scope.go:117] "RemoveContainer" containerID="fd08cb45146eab40252b2452a42e5aa9541c56dd4132484051f58a0cc9626b0c" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.517000 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-s5lm4"] Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.554229 4787 scope.go:117] "RemoveContainer" containerID="2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28" Dec 03 18:40:36 crc kubenswrapper[4787]: E1203 18:40:36.554624 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28\": container with ID starting with 2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28 not found: ID does not exist" containerID="2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.554668 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28"} err="failed to get container status \"2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28\": rpc error: code = NotFound desc = could not find container \"2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28\": container with ID starting with 2fa436cd36f52edce137acebca0f289aeaac211a5f78c8adb8ccd99a0cde1e28 not found: ID does not exist" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.554691 4787 scope.go:117] "RemoveContainer" containerID="25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727" Dec 03 18:40:36 crc kubenswrapper[4787]: E1203 18:40:36.555266 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727\": container with ID starting with 25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727 not found: ID does not exist" containerID="25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.555300 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727"} err="failed to get container status \"25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727\": rpc error: code = NotFound desc = could not find container \"25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727\": container with ID starting with 25ce3c68a8126761cace8de567c0d5706b2a5e71165c19eede8d8ef84d1a4727 not found: ID does not exist" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.555318 4787 scope.go:117] "RemoveContainer" containerID="fd08cb45146eab40252b2452a42e5aa9541c56dd4132484051f58a0cc9626b0c" Dec 03 18:40:36 crc kubenswrapper[4787]: E1203 18:40:36.555563 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd08cb45146eab40252b2452a42e5aa9541c56dd4132484051f58a0cc9626b0c\": container with ID starting with fd08cb45146eab40252b2452a42e5aa9541c56dd4132484051f58a0cc9626b0c not found: ID does not exist" containerID="fd08cb45146eab40252b2452a42e5aa9541c56dd4132484051f58a0cc9626b0c" Dec 03 18:40:36 crc kubenswrapper[4787]: I1203 18:40:36.555596 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd08cb45146eab40252b2452a42e5aa9541c56dd4132484051f58a0cc9626b0c"} err="failed to get container status \"fd08cb45146eab40252b2452a42e5aa9541c56dd4132484051f58a0cc9626b0c\": rpc error: code = NotFound desc = could not find container \"fd08cb45146eab40252b2452a42e5aa9541c56dd4132484051f58a0cc9626b0c\": container with ID starting with fd08cb45146eab40252b2452a42e5aa9541c56dd4132484051f58a0cc9626b0c not found: ID does not exist" Dec 03 18:40:37 crc kubenswrapper[4787]: I1203 18:40:37.782945 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" path="/var/lib/kubelet/pods/afa83387-b414-4fcf-9a8a-1ec08ef8d6ff/volumes" Dec 03 18:40:46 crc kubenswrapper[4787]: I1203 18:40:46.062706 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-2dp4k"] Dec 03 18:40:46 crc kubenswrapper[4787]: I1203 18:40:46.076522 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-6db2-account-create-update-r8jpd"] Dec 03 18:40:46 crc kubenswrapper[4787]: I1203 18:40:46.086499 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-6db2-account-create-update-r8jpd"] Dec 03 18:40:46 crc kubenswrapper[4787]: I1203 18:40:46.095434 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-2dp4k"] Dec 03 18:40:47 crc kubenswrapper[4787]: I1203 18:40:47.780814 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3368ce70-bad5-4535-a666-ca5b751967d0" path="/var/lib/kubelet/pods/3368ce70-bad5-4535-a666-ca5b751967d0/volumes" Dec 03 18:40:47 crc kubenswrapper[4787]: I1203 18:40:47.782583 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff52f6c6-c90d-43e7-a0cf-3fb81534f2de" path="/var/lib/kubelet/pods/ff52f6c6-c90d-43e7-a0cf-3fb81534f2de/volumes" Dec 03 18:40:48 crc kubenswrapper[4787]: I1203 18:40:48.990345 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:40:48 crc kubenswrapper[4787]: I1203 18:40:48.990420 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:41:13 crc kubenswrapper[4787]: I1203 18:41:13.052447 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-bfb6h"] Dec 03 18:41:13 crc kubenswrapper[4787]: I1203 18:41:13.063443 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-bfb6h"] Dec 03 18:41:13 crc kubenswrapper[4787]: I1203 18:41:13.783164 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b8e3c8d9-6215-4247-a39a-27a99e53a33f" path="/var/lib/kubelet/pods/b8e3c8d9-6215-4247-a39a-27a99e53a33f/volumes" Dec 03 18:41:18 crc kubenswrapper[4787]: I1203 18:41:18.989423 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:41:18 crc kubenswrapper[4787]: I1203 18:41:18.990093 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:41:18 crc kubenswrapper[4787]: I1203 18:41:18.990168 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 18:41:18 crc kubenswrapper[4787]: I1203 18:41:18.991706 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:41:18 crc kubenswrapper[4787]: I1203 18:41:18.991793 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" gracePeriod=600 Dec 03 18:41:19 crc kubenswrapper[4787]: E1203 18:41:19.119480 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:41:20 crc kubenswrapper[4787]: I1203 18:41:20.010594 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" exitCode=0 Dec 03 18:41:20 crc kubenswrapper[4787]: I1203 18:41:20.010681 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291"} Dec 03 18:41:20 crc kubenswrapper[4787]: I1203 18:41:20.010962 4787 scope.go:117] "RemoveContainer" containerID="743ebb9764dc004d652685c104dc6240ac82ed6fb6fb7bd45471642d78f70ab2" Dec 03 18:41:20 crc kubenswrapper[4787]: I1203 18:41:20.012163 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:41:20 crc kubenswrapper[4787]: E1203 18:41:20.012821 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:41:28 crc kubenswrapper[4787]: I1203 18:41:28.365933 4787 scope.go:117] "RemoveContainer" containerID="a62c7a1fd363768b8be82b42194d35c1e13c0e6afcbab83729a12e2d0f85cc0c" Dec 03 18:41:28 crc kubenswrapper[4787]: I1203 18:41:28.398774 4787 scope.go:117] "RemoveContainer" containerID="1b3628d89d8d0e7add80979b0ba64f0cb7fac1217227a14d6b843dcd250ddbc2" Dec 03 18:41:28 crc kubenswrapper[4787]: I1203 18:41:28.467087 4787 scope.go:117] "RemoveContainer" containerID="d0fa1c521211d166eac35c4a193398805fabcf828e8fae550e75dbde8806cc44" Dec 03 18:41:32 crc kubenswrapper[4787]: I1203 18:41:32.767922 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:41:32 crc kubenswrapper[4787]: E1203 18:41:32.769049 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:41:47 crc kubenswrapper[4787]: I1203 18:41:47.766490 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:41:47 crc kubenswrapper[4787]: E1203 18:41:47.767188 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:42:01 crc kubenswrapper[4787]: I1203 18:42:01.766491 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:42:01 crc kubenswrapper[4787]: E1203 18:42:01.767355 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:42:15 crc kubenswrapper[4787]: I1203 18:42:15.766575 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:42:15 crc kubenswrapper[4787]: E1203 18:42:15.767359 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:42:30 crc kubenswrapper[4787]: I1203 18:42:30.766577 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:42:30 crc kubenswrapper[4787]: E1203 18:42:30.767671 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:42:41 crc kubenswrapper[4787]: I1203 18:42:41.771193 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:42:41 crc kubenswrapper[4787]: E1203 18:42:41.772348 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:42:52 crc kubenswrapper[4787]: I1203 18:42:52.765521 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:42:52 crc kubenswrapper[4787]: E1203 18:42:52.766407 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:43:04 crc kubenswrapper[4787]: I1203 18:43:04.767061 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:43:04 crc kubenswrapper[4787]: E1203 18:43:04.768441 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:43:19 crc kubenswrapper[4787]: I1203 18:43:19.766551 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:43:19 crc kubenswrapper[4787]: E1203 18:43:19.767773 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:43:32 crc kubenswrapper[4787]: I1203 18:43:32.766357 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:43:32 crc kubenswrapper[4787]: E1203 18:43:32.767667 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:43:45 crc kubenswrapper[4787]: I1203 18:43:45.766437 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:43:45 crc kubenswrapper[4787]: E1203 18:43:45.767614 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:44:00 crc kubenswrapper[4787]: I1203 18:44:00.767069 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:44:00 crc kubenswrapper[4787]: E1203 18:44:00.768541 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.103889 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-mzx8x"] Dec 03 18:44:06 crc kubenswrapper[4787]: E1203 18:44:06.105610 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" containerName="extract-utilities" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.105629 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" containerName="extract-utilities" Dec 03 18:44:06 crc kubenswrapper[4787]: E1203 18:44:06.105650 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" containerName="registry-server" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.105658 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" containerName="registry-server" Dec 03 18:44:06 crc kubenswrapper[4787]: E1203 18:44:06.105668 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" containerName="extract-content" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.105675 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" containerName="extract-content" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.105947 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="afa83387-b414-4fcf-9a8a-1ec08ef8d6ff" containerName="registry-server" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.110268 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.125649 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mzx8x"] Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.230284 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q7qbk\" (UniqueName: \"kubernetes.io/projected/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-kube-api-access-q7qbk\") pod \"community-operators-mzx8x\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.230405 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-utilities\") pod \"community-operators-mzx8x\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.230446 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-catalog-content\") pod \"community-operators-mzx8x\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.332465 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-utilities\") pod \"community-operators-mzx8x\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.332554 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-catalog-content\") pod \"community-operators-mzx8x\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.332838 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q7qbk\" (UniqueName: \"kubernetes.io/projected/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-kube-api-access-q7qbk\") pod \"community-operators-mzx8x\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.333925 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-utilities\") pod \"community-operators-mzx8x\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.334256 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-catalog-content\") pod \"community-operators-mzx8x\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.354845 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q7qbk\" (UniqueName: \"kubernetes.io/projected/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-kube-api-access-q7qbk\") pod \"community-operators-mzx8x\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.433088 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:06 crc kubenswrapper[4787]: I1203 18:44:06.978337 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-mzx8x"] Dec 03 18:44:06 crc kubenswrapper[4787]: W1203 18:44:06.984510 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1cdcb520_4c76_4d5e_bc7b_71af32d351ed.slice/crio-8467f137cfa1326f97f29c7e492085b275afcef64286b49c9737966ab9cc3141 WatchSource:0}: Error finding container 8467f137cfa1326f97f29c7e492085b275afcef64286b49c9737966ab9cc3141: Status 404 returned error can't find the container with id 8467f137cfa1326f97f29c7e492085b275afcef64286b49c9737966ab9cc3141 Dec 03 18:44:07 crc kubenswrapper[4787]: I1203 18:44:07.181918 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzx8x" event={"ID":"1cdcb520-4c76-4d5e-bc7b-71af32d351ed","Type":"ContainerStarted","Data":"8467f137cfa1326f97f29c7e492085b275afcef64286b49c9737966ab9cc3141"} Dec 03 18:44:08 crc kubenswrapper[4787]: I1203 18:44:08.193183 4787 generic.go:334] "Generic (PLEG): container finished" podID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" containerID="1692d25820c267e99d627f76dffbfd489c6c9931d68669c60b8c1ce8ab7c96aa" exitCode=0 Dec 03 18:44:08 crc kubenswrapper[4787]: I1203 18:44:08.193239 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzx8x" event={"ID":"1cdcb520-4c76-4d5e-bc7b-71af32d351ed","Type":"ContainerDied","Data":"1692d25820c267e99d627f76dffbfd489c6c9931d68669c60b8c1ce8ab7c96aa"} Dec 03 18:44:09 crc kubenswrapper[4787]: I1203 18:44:09.210631 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzx8x" event={"ID":"1cdcb520-4c76-4d5e-bc7b-71af32d351ed","Type":"ContainerStarted","Data":"316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46"} Dec 03 18:44:10 crc kubenswrapper[4787]: I1203 18:44:10.230475 4787 generic.go:334] "Generic (PLEG): container finished" podID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" containerID="316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46" exitCode=0 Dec 03 18:44:10 crc kubenswrapper[4787]: I1203 18:44:10.230599 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzx8x" event={"ID":"1cdcb520-4c76-4d5e-bc7b-71af32d351ed","Type":"ContainerDied","Data":"316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46"} Dec 03 18:44:12 crc kubenswrapper[4787]: I1203 18:44:12.255834 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzx8x" event={"ID":"1cdcb520-4c76-4d5e-bc7b-71af32d351ed","Type":"ContainerStarted","Data":"1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93"} Dec 03 18:44:12 crc kubenswrapper[4787]: I1203 18:44:12.294494 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-mzx8x" podStartSLOduration=3.83616499 podStartE2EDuration="6.294465613s" podCreationTimestamp="2025-12-03 18:44:06 +0000 UTC" firstStartedPulling="2025-12-03 18:44:08.197769455 +0000 UTC m=+5465.015240414" lastFinishedPulling="2025-12-03 18:44:10.656070078 +0000 UTC m=+5467.473541037" observedRunningTime="2025-12-03 18:44:12.277525561 +0000 UTC m=+5469.094996550" watchObservedRunningTime="2025-12-03 18:44:12.294465613 +0000 UTC m=+5469.111936602" Dec 03 18:44:15 crc kubenswrapper[4787]: I1203 18:44:15.766906 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:44:15 crc kubenswrapper[4787]: E1203 18:44:15.767985 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:44:16 crc kubenswrapper[4787]: I1203 18:44:16.434609 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:16 crc kubenswrapper[4787]: I1203 18:44:16.435317 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:16 crc kubenswrapper[4787]: I1203 18:44:16.490651 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:17 crc kubenswrapper[4787]: I1203 18:44:17.425588 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:17 crc kubenswrapper[4787]: I1203 18:44:17.476670 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mzx8x"] Dec 03 18:44:19 crc kubenswrapper[4787]: I1203 18:44:19.393401 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-mzx8x" podUID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" containerName="registry-server" containerID="cri-o://1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93" gracePeriod=2 Dec 03 18:44:19 crc kubenswrapper[4787]: I1203 18:44:19.990932 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.095557 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q7qbk\" (UniqueName: \"kubernetes.io/projected/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-kube-api-access-q7qbk\") pod \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.095849 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-catalog-content\") pod \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.096003 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-utilities\") pod \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\" (UID: \"1cdcb520-4c76-4d5e-bc7b-71af32d351ed\") " Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.096580 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-utilities" (OuterVolumeSpecName: "utilities") pod "1cdcb520-4c76-4d5e-bc7b-71af32d351ed" (UID: "1cdcb520-4c76-4d5e-bc7b-71af32d351ed"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.107325 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-kube-api-access-q7qbk" (OuterVolumeSpecName: "kube-api-access-q7qbk") pod "1cdcb520-4c76-4d5e-bc7b-71af32d351ed" (UID: "1cdcb520-4c76-4d5e-bc7b-71af32d351ed"). InnerVolumeSpecName "kube-api-access-q7qbk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.142309 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1cdcb520-4c76-4d5e-bc7b-71af32d351ed" (UID: "1cdcb520-4c76-4d5e-bc7b-71af32d351ed"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.198686 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.198718 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.198729 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q7qbk\" (UniqueName: \"kubernetes.io/projected/1cdcb520-4c76-4d5e-bc7b-71af32d351ed-kube-api-access-q7qbk\") on node \"crc\" DevicePath \"\"" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.403530 4787 generic.go:334] "Generic (PLEG): container finished" podID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" containerID="1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93" exitCode=0 Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.403581 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzx8x" event={"ID":"1cdcb520-4c76-4d5e-bc7b-71af32d351ed","Type":"ContainerDied","Data":"1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93"} Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.403606 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-mzx8x" event={"ID":"1cdcb520-4c76-4d5e-bc7b-71af32d351ed","Type":"ContainerDied","Data":"8467f137cfa1326f97f29c7e492085b275afcef64286b49c9737966ab9cc3141"} Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.403625 4787 scope.go:117] "RemoveContainer" containerID="1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.403743 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-mzx8x" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.449959 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-mzx8x"] Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.450071 4787 scope.go:117] "RemoveContainer" containerID="316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.463538 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-mzx8x"] Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.480663 4787 scope.go:117] "RemoveContainer" containerID="1692d25820c267e99d627f76dffbfd489c6c9931d68669c60b8c1ce8ab7c96aa" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.530950 4787 scope.go:117] "RemoveContainer" containerID="1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93" Dec 03 18:44:20 crc kubenswrapper[4787]: E1203 18:44:20.531818 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93\": container with ID starting with 1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93 not found: ID does not exist" containerID="1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.531859 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93"} err="failed to get container status \"1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93\": rpc error: code = NotFound desc = could not find container \"1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93\": container with ID starting with 1bf60a9448ff7cf3b65514d52f2001dc937269804c6ad51df0c140f2c2ac3f93 not found: ID does not exist" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.531885 4787 scope.go:117] "RemoveContainer" containerID="316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46" Dec 03 18:44:20 crc kubenswrapper[4787]: E1203 18:44:20.532403 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46\": container with ID starting with 316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46 not found: ID does not exist" containerID="316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.532472 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46"} err="failed to get container status \"316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46\": rpc error: code = NotFound desc = could not find container \"316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46\": container with ID starting with 316589a606035884281ca2116f4cb97743d663eef1b2363cf07d198622fe4c46 not found: ID does not exist" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.532499 4787 scope.go:117] "RemoveContainer" containerID="1692d25820c267e99d627f76dffbfd489c6c9931d68669c60b8c1ce8ab7c96aa" Dec 03 18:44:20 crc kubenswrapper[4787]: E1203 18:44:20.532951 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1692d25820c267e99d627f76dffbfd489c6c9931d68669c60b8c1ce8ab7c96aa\": container with ID starting with 1692d25820c267e99d627f76dffbfd489c6c9931d68669c60b8c1ce8ab7c96aa not found: ID does not exist" containerID="1692d25820c267e99d627f76dffbfd489c6c9931d68669c60b8c1ce8ab7c96aa" Dec 03 18:44:20 crc kubenswrapper[4787]: I1203 18:44:20.533460 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1692d25820c267e99d627f76dffbfd489c6c9931d68669c60b8c1ce8ab7c96aa"} err="failed to get container status \"1692d25820c267e99d627f76dffbfd489c6c9931d68669c60b8c1ce8ab7c96aa\": rpc error: code = NotFound desc = could not find container \"1692d25820c267e99d627f76dffbfd489c6c9931d68669c60b8c1ce8ab7c96aa\": container with ID starting with 1692d25820c267e99d627f76dffbfd489c6c9931d68669c60b8c1ce8ab7c96aa not found: ID does not exist" Dec 03 18:44:21 crc kubenswrapper[4787]: I1203 18:44:21.791237 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" path="/var/lib/kubelet/pods/1cdcb520-4c76-4d5e-bc7b-71af32d351ed/volumes" Dec 03 18:44:29 crc kubenswrapper[4787]: I1203 18:44:29.766662 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:44:29 crc kubenswrapper[4787]: E1203 18:44:29.767777 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.002668 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-j95ht"] Dec 03 18:44:32 crc kubenswrapper[4787]: E1203 18:44:32.003716 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" containerName="extract-utilities" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.003741 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" containerName="extract-utilities" Dec 03 18:44:32 crc kubenswrapper[4787]: E1203 18:44:32.003777 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" containerName="extract-content" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.003790 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" containerName="extract-content" Dec 03 18:44:32 crc kubenswrapper[4787]: E1203 18:44:32.003834 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" containerName="registry-server" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.003848 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" containerName="registry-server" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.004292 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1cdcb520-4c76-4d5e-bc7b-71af32d351ed" containerName="registry-server" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.007371 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.037488 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j95ht"] Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.100532 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xdc6\" (UniqueName: \"kubernetes.io/projected/3efdf95a-b018-4550-9853-61869474de08-kube-api-access-8xdc6\") pod \"redhat-operators-j95ht\" (UID: \"3efdf95a-b018-4550-9853-61869474de08\") " pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.100696 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efdf95a-b018-4550-9853-61869474de08-catalog-content\") pod \"redhat-operators-j95ht\" (UID: \"3efdf95a-b018-4550-9853-61869474de08\") " pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.101009 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efdf95a-b018-4550-9853-61869474de08-utilities\") pod \"redhat-operators-j95ht\" (UID: \"3efdf95a-b018-4550-9853-61869474de08\") " pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.203321 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efdf95a-b018-4550-9853-61869474de08-utilities\") pod \"redhat-operators-j95ht\" (UID: \"3efdf95a-b018-4550-9853-61869474de08\") " pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.203529 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xdc6\" (UniqueName: \"kubernetes.io/projected/3efdf95a-b018-4550-9853-61869474de08-kube-api-access-8xdc6\") pod \"redhat-operators-j95ht\" (UID: \"3efdf95a-b018-4550-9853-61869474de08\") " pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.203562 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efdf95a-b018-4550-9853-61869474de08-catalog-content\") pod \"redhat-operators-j95ht\" (UID: \"3efdf95a-b018-4550-9853-61869474de08\") " pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.204047 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3efdf95a-b018-4550-9853-61869474de08-catalog-content\") pod \"redhat-operators-j95ht\" (UID: \"3efdf95a-b018-4550-9853-61869474de08\") " pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.204228 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3efdf95a-b018-4550-9853-61869474de08-utilities\") pod \"redhat-operators-j95ht\" (UID: \"3efdf95a-b018-4550-9853-61869474de08\") " pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.226582 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xdc6\" (UniqueName: \"kubernetes.io/projected/3efdf95a-b018-4550-9853-61869474de08-kube-api-access-8xdc6\") pod \"redhat-operators-j95ht\" (UID: \"3efdf95a-b018-4550-9853-61869474de08\") " pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.329042 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:32 crc kubenswrapper[4787]: I1203 18:44:32.821366 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j95ht"] Dec 03 18:44:33 crc kubenswrapper[4787]: I1203 18:44:33.577578 4787 generic.go:334] "Generic (PLEG): container finished" podID="3efdf95a-b018-4550-9853-61869474de08" containerID="16f94e4e5633acc162b82857f7420513295e882957b2bc7c3c6dca9c2571c215" exitCode=0 Dec 03 18:44:33 crc kubenswrapper[4787]: I1203 18:44:33.577830 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j95ht" event={"ID":"3efdf95a-b018-4550-9853-61869474de08","Type":"ContainerDied","Data":"16f94e4e5633acc162b82857f7420513295e882957b2bc7c3c6dca9c2571c215"} Dec 03 18:44:33 crc kubenswrapper[4787]: I1203 18:44:33.577855 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j95ht" event={"ID":"3efdf95a-b018-4550-9853-61869474de08","Type":"ContainerStarted","Data":"562314a6425041ea1a7f49cde482601152156efa8fc4cc63be1ccf95185e0153"} Dec 03 18:44:41 crc kubenswrapper[4787]: I1203 18:44:41.768166 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:44:41 crc kubenswrapper[4787]: E1203 18:44:41.769841 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:44:42 crc kubenswrapper[4787]: I1203 18:44:42.682461 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j95ht" event={"ID":"3efdf95a-b018-4550-9853-61869474de08","Type":"ContainerStarted","Data":"e0a9077e83dde1d5a27f2584a317653900205432581889dc4b01b2c1c32916e4"} Dec 03 18:44:46 crc kubenswrapper[4787]: I1203 18:44:46.730009 4787 generic.go:334] "Generic (PLEG): container finished" podID="3efdf95a-b018-4550-9853-61869474de08" containerID="e0a9077e83dde1d5a27f2584a317653900205432581889dc4b01b2c1c32916e4" exitCode=0 Dec 03 18:44:46 crc kubenswrapper[4787]: I1203 18:44:46.730190 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j95ht" event={"ID":"3efdf95a-b018-4550-9853-61869474de08","Type":"ContainerDied","Data":"e0a9077e83dde1d5a27f2584a317653900205432581889dc4b01b2c1c32916e4"} Dec 03 18:44:47 crc kubenswrapper[4787]: I1203 18:44:47.742535 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-j95ht" event={"ID":"3efdf95a-b018-4550-9853-61869474de08","Type":"ContainerStarted","Data":"2518e390ffbf893c1f424eb4ae0a3ac19a4863bcad2fa53bde0de12e56a1aaa4"} Dec 03 18:44:47 crc kubenswrapper[4787]: I1203 18:44:47.763402 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-j95ht" podStartSLOduration=3.093719654 podStartE2EDuration="16.763385769s" podCreationTimestamp="2025-12-03 18:44:31 +0000 UTC" firstStartedPulling="2025-12-03 18:44:33.580057318 +0000 UTC m=+5490.397528277" lastFinishedPulling="2025-12-03 18:44:47.249723403 +0000 UTC m=+5504.067194392" observedRunningTime="2025-12-03 18:44:47.759992439 +0000 UTC m=+5504.577463398" watchObservedRunningTime="2025-12-03 18:44:47.763385769 +0000 UTC m=+5504.580856718" Dec 03 18:44:52 crc kubenswrapper[4787]: I1203 18:44:52.329487 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:52 crc kubenswrapper[4787]: I1203 18:44:52.330105 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:44:53 crc kubenswrapper[4787]: I1203 18:44:53.384864 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-j95ht" podUID="3efdf95a-b018-4550-9853-61869474de08" containerName="registry-server" probeResult="failure" output=< Dec 03 18:44:53 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 18:44:53 crc kubenswrapper[4787]: > Dec 03 18:44:54 crc kubenswrapper[4787]: I1203 18:44:54.766000 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:44:54 crc kubenswrapper[4787]: E1203 18:44:54.766576 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.176734 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn"] Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.179828 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.185457 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.185933 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.191592 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn"] Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.274269 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-secret-volume\") pod \"collect-profiles-29413125-27rjn\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.274420 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9glc9\" (UniqueName: \"kubernetes.io/projected/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-kube-api-access-9glc9\") pod \"collect-profiles-29413125-27rjn\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.274614 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-config-volume\") pod \"collect-profiles-29413125-27rjn\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.376931 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-secret-volume\") pod \"collect-profiles-29413125-27rjn\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.377042 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9glc9\" (UniqueName: \"kubernetes.io/projected/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-kube-api-access-9glc9\") pod \"collect-profiles-29413125-27rjn\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.377137 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-config-volume\") pod \"collect-profiles-29413125-27rjn\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.378229 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-config-volume\") pod \"collect-profiles-29413125-27rjn\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.386854 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-secret-volume\") pod \"collect-profiles-29413125-27rjn\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.397670 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9glc9\" (UniqueName: \"kubernetes.io/projected/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-kube-api-access-9glc9\") pod \"collect-profiles-29413125-27rjn\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:00 crc kubenswrapper[4787]: I1203 18:45:00.514147 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:01 crc kubenswrapper[4787]: I1203 18:45:01.119455 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn"] Dec 03 18:45:01 crc kubenswrapper[4787]: I1203 18:45:01.919439 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" event={"ID":"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc","Type":"ContainerStarted","Data":"ad4c6b273d25e555d7adc0ad015f11daa6508e6bf741621e5b1de7c5a0d0f9ce"} Dec 03 18:45:01 crc kubenswrapper[4787]: I1203 18:45:01.919755 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" event={"ID":"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc","Type":"ContainerStarted","Data":"2f19a6fe10be5e9e381fffba76a2322b0fa3d28142f3ce25d63d1fd083872edf"} Dec 03 18:45:01 crc kubenswrapper[4787]: I1203 18:45:01.951003 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" podStartSLOduration=1.950983124 podStartE2EDuration="1.950983124s" podCreationTimestamp="2025-12-03 18:45:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 18:45:01.943562056 +0000 UTC m=+5518.761033055" watchObservedRunningTime="2025-12-03 18:45:01.950983124 +0000 UTC m=+5518.768454083" Dec 03 18:45:02 crc kubenswrapper[4787]: I1203 18:45:02.691584 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:45:02 crc kubenswrapper[4787]: I1203 18:45:02.768524 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-j95ht" Dec 03 18:45:02 crc kubenswrapper[4787]: I1203 18:45:02.933815 4787 generic.go:334] "Generic (PLEG): container finished" podID="862bfb78-9edb-409f-bdca-9dcc2f6a3cdc" containerID="ad4c6b273d25e555d7adc0ad015f11daa6508e6bf741621e5b1de7c5a0d0f9ce" exitCode=0 Dec 03 18:45:02 crc kubenswrapper[4787]: I1203 18:45:02.933960 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" event={"ID":"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc","Type":"ContainerDied","Data":"ad4c6b273d25e555d7adc0ad015f11daa6508e6bf741621e5b1de7c5a0d0f9ce"} Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.026470 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-j95ht"] Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.199550 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rs8f5"] Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.199916 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-rs8f5" podUID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" containerName="registry-server" containerID="cri-o://d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04" gracePeriod=2 Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.666462 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.776026 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pgchs\" (UniqueName: \"kubernetes.io/projected/1b309950-19b2-4d34-b4b8-f5336e6c47f5-kube-api-access-pgchs\") pod \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.776130 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-catalog-content\") pod \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.776251 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-utilities\") pod \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\" (UID: \"1b309950-19b2-4d34-b4b8-f5336e6c47f5\") " Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.777933 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-utilities" (OuterVolumeSpecName: "utilities") pod "1b309950-19b2-4d34-b4b8-f5336e6c47f5" (UID: "1b309950-19b2-4d34-b4b8-f5336e6c47f5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.796853 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b309950-19b2-4d34-b4b8-f5336e6c47f5-kube-api-access-pgchs" (OuterVolumeSpecName: "kube-api-access-pgchs") pod "1b309950-19b2-4d34-b4b8-f5336e6c47f5" (UID: "1b309950-19b2-4d34-b4b8-f5336e6c47f5"). InnerVolumeSpecName "kube-api-access-pgchs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.879358 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pgchs\" (UniqueName: \"kubernetes.io/projected/1b309950-19b2-4d34-b4b8-f5336e6c47f5-kube-api-access-pgchs\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.879386 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.920189 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1b309950-19b2-4d34-b4b8-f5336e6c47f5" (UID: "1b309950-19b2-4d34-b4b8-f5336e6c47f5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.950049 4787 generic.go:334] "Generic (PLEG): container finished" podID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" containerID="d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04" exitCode=0 Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.951005 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-rs8f5" Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.951032 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs8f5" event={"ID":"1b309950-19b2-4d34-b4b8-f5336e6c47f5","Type":"ContainerDied","Data":"d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04"} Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.951086 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-rs8f5" event={"ID":"1b309950-19b2-4d34-b4b8-f5336e6c47f5","Type":"ContainerDied","Data":"46ca95fe5526b29f4eeed28d926509d673efe896d0684e4af3d33bdd47c1bce2"} Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.951105 4787 scope.go:117] "RemoveContainer" containerID="d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04" Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.981139 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1b309950-19b2-4d34-b4b8-f5336e6c47f5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.989178 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-rs8f5"] Dec 03 18:45:03 crc kubenswrapper[4787]: I1203 18:45:03.997560 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-rs8f5"] Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.006404 4787 scope.go:117] "RemoveContainer" containerID="2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.036330 4787 scope.go:117] "RemoveContainer" containerID="f382471f7933baadc81f1f42ca12c89554df5d2f997e70a6b9ebbceaedfc2e6e" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.079305 4787 scope.go:117] "RemoveContainer" containerID="d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04" Dec 03 18:45:04 crc kubenswrapper[4787]: E1203 18:45:04.079917 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04\": container with ID starting with d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04 not found: ID does not exist" containerID="d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.079961 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04"} err="failed to get container status \"d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04\": rpc error: code = NotFound desc = could not find container \"d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04\": container with ID starting with d93a93be3a72d12f83824133588d6fd8d48dca53df3b1ab620bd584afa297c04 not found: ID does not exist" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.079989 4787 scope.go:117] "RemoveContainer" containerID="2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776" Dec 03 18:45:04 crc kubenswrapper[4787]: E1203 18:45:04.080507 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776\": container with ID starting with 2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776 not found: ID does not exist" containerID="2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.080528 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776"} err="failed to get container status \"2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776\": rpc error: code = NotFound desc = could not find container \"2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776\": container with ID starting with 2a0385274f5d3fc8402dc3f3e2af19cb8966fbb6bd538515eeb6e90ac5783776 not found: ID does not exist" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.080541 4787 scope.go:117] "RemoveContainer" containerID="f382471f7933baadc81f1f42ca12c89554df5d2f997e70a6b9ebbceaedfc2e6e" Dec 03 18:45:04 crc kubenswrapper[4787]: E1203 18:45:04.080761 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f382471f7933baadc81f1f42ca12c89554df5d2f997e70a6b9ebbceaedfc2e6e\": container with ID starting with f382471f7933baadc81f1f42ca12c89554df5d2f997e70a6b9ebbceaedfc2e6e not found: ID does not exist" containerID="f382471f7933baadc81f1f42ca12c89554df5d2f997e70a6b9ebbceaedfc2e6e" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.080781 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f382471f7933baadc81f1f42ca12c89554df5d2f997e70a6b9ebbceaedfc2e6e"} err="failed to get container status \"f382471f7933baadc81f1f42ca12c89554df5d2f997e70a6b9ebbceaedfc2e6e\": rpc error: code = NotFound desc = could not find container \"f382471f7933baadc81f1f42ca12c89554df5d2f997e70a6b9ebbceaedfc2e6e\": container with ID starting with f382471f7933baadc81f1f42ca12c89554df5d2f997e70a6b9ebbceaedfc2e6e not found: ID does not exist" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.259651 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.396932 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-secret-volume\") pod \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.396988 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-config-volume\") pod \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.397061 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9glc9\" (UniqueName: \"kubernetes.io/projected/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-kube-api-access-9glc9\") pod \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\" (UID: \"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc\") " Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.397484 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-config-volume" (OuterVolumeSpecName: "config-volume") pod "862bfb78-9edb-409f-bdca-9dcc2f6a3cdc" (UID: "862bfb78-9edb-409f-bdca-9dcc2f6a3cdc"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.398096 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.404130 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "862bfb78-9edb-409f-bdca-9dcc2f6a3cdc" (UID: "862bfb78-9edb-409f-bdca-9dcc2f6a3cdc"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.405400 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-kube-api-access-9glc9" (OuterVolumeSpecName: "kube-api-access-9glc9") pod "862bfb78-9edb-409f-bdca-9dcc2f6a3cdc" (UID: "862bfb78-9edb-409f-bdca-9dcc2f6a3cdc"). InnerVolumeSpecName "kube-api-access-9glc9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.499947 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.499987 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9glc9\" (UniqueName: \"kubernetes.io/projected/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc-kube-api-access-9glc9\") on node \"crc\" DevicePath \"\"" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.965363 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" event={"ID":"862bfb78-9edb-409f-bdca-9dcc2f6a3cdc","Type":"ContainerDied","Data":"2f19a6fe10be5e9e381fffba76a2322b0fa3d28142f3ce25d63d1fd083872edf"} Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.966466 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2f19a6fe10be5e9e381fffba76a2322b0fa3d28142f3ce25d63d1fd083872edf" Dec 03 18:45:04 crc kubenswrapper[4787]: I1203 18:45:04.965428 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn" Dec 03 18:45:05 crc kubenswrapper[4787]: I1203 18:45:05.030625 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2"] Dec 03 18:45:05 crc kubenswrapper[4787]: I1203 18:45:05.044219 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413080-9hmf2"] Dec 03 18:45:05 crc kubenswrapper[4787]: I1203 18:45:05.777812 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" path="/var/lib/kubelet/pods/1b309950-19b2-4d34-b4b8-f5336e6c47f5/volumes" Dec 03 18:45:05 crc kubenswrapper[4787]: I1203 18:45:05.779231 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f03a6f41-66de-4c21-84a1-8c30b66c80b8" path="/var/lib/kubelet/pods/f03a6f41-66de-4c21-84a1-8c30b66c80b8/volumes" Dec 03 18:45:07 crc kubenswrapper[4787]: I1203 18:45:07.766384 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:45:07 crc kubenswrapper[4787]: E1203 18:45:07.766862 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:45:18 crc kubenswrapper[4787]: I1203 18:45:18.766816 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:45:18 crc kubenswrapper[4787]: E1203 18:45:18.768167 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:45:28 crc kubenswrapper[4787]: I1203 18:45:28.701337 4787 scope.go:117] "RemoveContainer" containerID="14867a714d1ee66f92e7c58dd907908da0d9d440efc984f967c8903efd0addf3" Dec 03 18:45:30 crc kubenswrapper[4787]: I1203 18:45:30.766110 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:45:30 crc kubenswrapper[4787]: E1203 18:45:30.767049 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:45:42 crc kubenswrapper[4787]: I1203 18:45:42.765821 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:45:42 crc kubenswrapper[4787]: E1203 18:45:42.767950 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:45:55 crc kubenswrapper[4787]: I1203 18:45:55.766969 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:45:55 crc kubenswrapper[4787]: E1203 18:45:55.767825 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:46:06 crc kubenswrapper[4787]: I1203 18:46:06.766078 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:46:06 crc kubenswrapper[4787]: E1203 18:46:06.766868 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:46:18 crc kubenswrapper[4787]: I1203 18:46:18.768249 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:46:18 crc kubenswrapper[4787]: E1203 18:46:18.769694 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:46:29 crc kubenswrapper[4787]: I1203 18:46:29.771099 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:46:31 crc kubenswrapper[4787]: I1203 18:46:31.052163 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"710bb39b7b391e5d3782590ada5ae767a151048d56b184accb8ea5caf20394fa"} Dec 03 18:48:48 crc kubenswrapper[4787]: I1203 18:48:48.989768 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:48:48 crc kubenswrapper[4787]: I1203 18:48:48.990277 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.322145 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-69rx8"] Dec 03 18:49:12 crc kubenswrapper[4787]: E1203 18:49:12.323400 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="862bfb78-9edb-409f-bdca-9dcc2f6a3cdc" containerName="collect-profiles" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.323421 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="862bfb78-9edb-409f-bdca-9dcc2f6a3cdc" containerName="collect-profiles" Dec 03 18:49:12 crc kubenswrapper[4787]: E1203 18:49:12.323456 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" containerName="registry-server" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.323468 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" containerName="registry-server" Dec 03 18:49:12 crc kubenswrapper[4787]: E1203 18:49:12.323538 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" containerName="extract-content" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.323553 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" containerName="extract-content" Dec 03 18:49:12 crc kubenswrapper[4787]: E1203 18:49:12.323575 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" containerName="extract-utilities" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.323587 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" containerName="extract-utilities" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.323963 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b309950-19b2-4d34-b4b8-f5336e6c47f5" containerName="registry-server" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.324056 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="862bfb78-9edb-409f-bdca-9dcc2f6a3cdc" containerName="collect-profiles" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.326947 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.338503 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-69rx8"] Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.359180 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-catalog-content\") pod \"redhat-marketplace-69rx8\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.359357 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-utilities\") pod \"redhat-marketplace-69rx8\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.359481 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79hgl\" (UniqueName: \"kubernetes.io/projected/7d4006f5-f23a-456a-a376-3ee6fe08a922-kube-api-access-79hgl\") pod \"redhat-marketplace-69rx8\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.465850 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-utilities\") pod \"redhat-marketplace-69rx8\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.465971 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79hgl\" (UniqueName: \"kubernetes.io/projected/7d4006f5-f23a-456a-a376-3ee6fe08a922-kube-api-access-79hgl\") pod \"redhat-marketplace-69rx8\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.466181 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-catalog-content\") pod \"redhat-marketplace-69rx8\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.466677 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-utilities\") pod \"redhat-marketplace-69rx8\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.466750 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-catalog-content\") pod \"redhat-marketplace-69rx8\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.484982 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79hgl\" (UniqueName: \"kubernetes.io/projected/7d4006f5-f23a-456a-a376-3ee6fe08a922-kube-api-access-79hgl\") pod \"redhat-marketplace-69rx8\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:12 crc kubenswrapper[4787]: I1203 18:49:12.671985 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:13 crc kubenswrapper[4787]: I1203 18:49:13.199751 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-69rx8"] Dec 03 18:49:13 crc kubenswrapper[4787]: I1203 18:49:13.343160 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69rx8" event={"ID":"7d4006f5-f23a-456a-a376-3ee6fe08a922","Type":"ContainerStarted","Data":"ad32bd31bf8a3f886f0e4cae34f1063f2009958fbde5a612287f199a082e6d98"} Dec 03 18:49:14 crc kubenswrapper[4787]: I1203 18:49:14.354403 4787 generic.go:334] "Generic (PLEG): container finished" podID="7d4006f5-f23a-456a-a376-3ee6fe08a922" containerID="ff155366710d511936086c001969cd32cc67a2173a4851d62a9d58555d471b0d" exitCode=0 Dec 03 18:49:14 crc kubenswrapper[4787]: I1203 18:49:14.354445 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69rx8" event={"ID":"7d4006f5-f23a-456a-a376-3ee6fe08a922","Type":"ContainerDied","Data":"ff155366710d511936086c001969cd32cc67a2173a4851d62a9d58555d471b0d"} Dec 03 18:49:14 crc kubenswrapper[4787]: I1203 18:49:14.357772 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:49:16 crc kubenswrapper[4787]: I1203 18:49:16.381467 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69rx8" event={"ID":"7d4006f5-f23a-456a-a376-3ee6fe08a922","Type":"ContainerStarted","Data":"257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7"} Dec 03 18:49:17 crc kubenswrapper[4787]: I1203 18:49:17.395501 4787 generic.go:334] "Generic (PLEG): container finished" podID="7d4006f5-f23a-456a-a376-3ee6fe08a922" containerID="257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7" exitCode=0 Dec 03 18:49:17 crc kubenswrapper[4787]: I1203 18:49:17.395628 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69rx8" event={"ID":"7d4006f5-f23a-456a-a376-3ee6fe08a922","Type":"ContainerDied","Data":"257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7"} Dec 03 18:49:18 crc kubenswrapper[4787]: I1203 18:49:18.412524 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69rx8" event={"ID":"7d4006f5-f23a-456a-a376-3ee6fe08a922","Type":"ContainerStarted","Data":"85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa"} Dec 03 18:49:18 crc kubenswrapper[4787]: I1203 18:49:18.989968 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:49:18 crc kubenswrapper[4787]: I1203 18:49:18.990058 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:49:22 crc kubenswrapper[4787]: I1203 18:49:22.672613 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:22 crc kubenswrapper[4787]: I1203 18:49:22.673198 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:23 crc kubenswrapper[4787]: I1203 18:49:23.561661 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:23 crc kubenswrapper[4787]: I1203 18:49:23.588028 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-69rx8" podStartSLOduration=7.919223658 podStartE2EDuration="11.587994099s" podCreationTimestamp="2025-12-03 18:49:12 +0000 UTC" firstStartedPulling="2025-12-03 18:49:14.357549012 +0000 UTC m=+5771.175019971" lastFinishedPulling="2025-12-03 18:49:18.026319463 +0000 UTC m=+5774.843790412" observedRunningTime="2025-12-03 18:49:18.434529764 +0000 UTC m=+5775.252000723" watchObservedRunningTime="2025-12-03 18:49:23.587994099 +0000 UTC m=+5780.405465068" Dec 03 18:49:23 crc kubenswrapper[4787]: I1203 18:49:23.627721 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:26 crc kubenswrapper[4787]: I1203 18:49:26.887829 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-69rx8"] Dec 03 18:49:26 crc kubenswrapper[4787]: I1203 18:49:26.889271 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-69rx8" podUID="7d4006f5-f23a-456a-a376-3ee6fe08a922" containerName="registry-server" containerID="cri-o://85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa" gracePeriod=2 Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.488926 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.528088 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-utilities\") pod \"7d4006f5-f23a-456a-a376-3ee6fe08a922\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.528681 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79hgl\" (UniqueName: \"kubernetes.io/projected/7d4006f5-f23a-456a-a376-3ee6fe08a922-kube-api-access-79hgl\") pod \"7d4006f5-f23a-456a-a376-3ee6fe08a922\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.528723 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-catalog-content\") pod \"7d4006f5-f23a-456a-a376-3ee6fe08a922\" (UID: \"7d4006f5-f23a-456a-a376-3ee6fe08a922\") " Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.530238 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-utilities" (OuterVolumeSpecName: "utilities") pod "7d4006f5-f23a-456a-a376-3ee6fe08a922" (UID: "7d4006f5-f23a-456a-a376-3ee6fe08a922"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.540600 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d4006f5-f23a-456a-a376-3ee6fe08a922-kube-api-access-79hgl" (OuterVolumeSpecName: "kube-api-access-79hgl") pod "7d4006f5-f23a-456a-a376-3ee6fe08a922" (UID: "7d4006f5-f23a-456a-a376-3ee6fe08a922"). InnerVolumeSpecName "kube-api-access-79hgl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.542792 4787 generic.go:334] "Generic (PLEG): container finished" podID="7d4006f5-f23a-456a-a376-3ee6fe08a922" containerID="85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa" exitCode=0 Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.542832 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69rx8" event={"ID":"7d4006f5-f23a-456a-a376-3ee6fe08a922","Type":"ContainerDied","Data":"85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa"} Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.542858 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-69rx8" event={"ID":"7d4006f5-f23a-456a-a376-3ee6fe08a922","Type":"ContainerDied","Data":"ad32bd31bf8a3f886f0e4cae34f1063f2009958fbde5a612287f199a082e6d98"} Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.542877 4787 scope.go:117] "RemoveContainer" containerID="85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.543006 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-69rx8" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.559157 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7d4006f5-f23a-456a-a376-3ee6fe08a922" (UID: "7d4006f5-f23a-456a-a376-3ee6fe08a922"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.611815 4787 scope.go:117] "RemoveContainer" containerID="257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.632182 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.632217 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79hgl\" (UniqueName: \"kubernetes.io/projected/7d4006f5-f23a-456a-a376-3ee6fe08a922-kube-api-access-79hgl\") on node \"crc\" DevicePath \"\"" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.632229 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7d4006f5-f23a-456a-a376-3ee6fe08a922-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.646227 4787 scope.go:117] "RemoveContainer" containerID="ff155366710d511936086c001969cd32cc67a2173a4851d62a9d58555d471b0d" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.740225 4787 scope.go:117] "RemoveContainer" containerID="85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa" Dec 03 18:49:27 crc kubenswrapper[4787]: E1203 18:49:27.740607 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa\": container with ID starting with 85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa not found: ID does not exist" containerID="85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.740642 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa"} err="failed to get container status \"85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa\": rpc error: code = NotFound desc = could not find container \"85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa\": container with ID starting with 85bb32a53c81445d0cafc020bd986e1e932d8e42ddf85e4807e2a04c9973a0fa not found: ID does not exist" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.740663 4787 scope.go:117] "RemoveContainer" containerID="257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7" Dec 03 18:49:27 crc kubenswrapper[4787]: E1203 18:49:27.741124 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7\": container with ID starting with 257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7 not found: ID does not exist" containerID="257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.741417 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7"} err="failed to get container status \"257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7\": rpc error: code = NotFound desc = could not find container \"257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7\": container with ID starting with 257e571a4cca3cc4c75133fc1e6cb2f04343137e59ea4c4ab3d25fd91ec07cd7 not found: ID does not exist" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.741442 4787 scope.go:117] "RemoveContainer" containerID="ff155366710d511936086c001969cd32cc67a2173a4851d62a9d58555d471b0d" Dec 03 18:49:27 crc kubenswrapper[4787]: E1203 18:49:27.741696 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff155366710d511936086c001969cd32cc67a2173a4851d62a9d58555d471b0d\": container with ID starting with ff155366710d511936086c001969cd32cc67a2173a4851d62a9d58555d471b0d not found: ID does not exist" containerID="ff155366710d511936086c001969cd32cc67a2173a4851d62a9d58555d471b0d" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.741720 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff155366710d511936086c001969cd32cc67a2173a4851d62a9d58555d471b0d"} err="failed to get container status \"ff155366710d511936086c001969cd32cc67a2173a4851d62a9d58555d471b0d\": rpc error: code = NotFound desc = could not find container \"ff155366710d511936086c001969cd32cc67a2173a4851d62a9d58555d471b0d\": container with ID starting with ff155366710d511936086c001969cd32cc67a2173a4851d62a9d58555d471b0d not found: ID does not exist" Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.872785 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-69rx8"] Dec 03 18:49:27 crc kubenswrapper[4787]: I1203 18:49:27.883067 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-69rx8"] Dec 03 18:49:29 crc kubenswrapper[4787]: I1203 18:49:29.787681 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d4006f5-f23a-456a-a376-3ee6fe08a922" path="/var/lib/kubelet/pods/7d4006f5-f23a-456a-a376-3ee6fe08a922/volumes" Dec 03 18:49:48 crc kubenswrapper[4787]: I1203 18:49:48.990057 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:49:48 crc kubenswrapper[4787]: I1203 18:49:48.990639 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:49:48 crc kubenswrapper[4787]: I1203 18:49:48.990692 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 18:49:48 crc kubenswrapper[4787]: I1203 18:49:48.991668 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"710bb39b7b391e5d3782590ada5ae767a151048d56b184accb8ea5caf20394fa"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:49:48 crc kubenswrapper[4787]: I1203 18:49:48.991734 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://710bb39b7b391e5d3782590ada5ae767a151048d56b184accb8ea5caf20394fa" gracePeriod=600 Dec 03 18:49:49 crc kubenswrapper[4787]: I1203 18:49:49.848798 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="710bb39b7b391e5d3782590ada5ae767a151048d56b184accb8ea5caf20394fa" exitCode=0 Dec 03 18:49:49 crc kubenswrapper[4787]: I1203 18:49:49.849485 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"710bb39b7b391e5d3782590ada5ae767a151048d56b184accb8ea5caf20394fa"} Dec 03 18:49:49 crc kubenswrapper[4787]: I1203 18:49:49.849512 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5"} Dec 03 18:49:49 crc kubenswrapper[4787]: I1203 18:49:49.849528 4787 scope.go:117] "RemoveContainer" containerID="15e5bb8fae4a7fbac72b42da08ad5affd47e2bd46604718f5a2844bb8f8e8291" Dec 03 18:50:22 crc kubenswrapper[4787]: I1203 18:50:22.803224 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-78pj2"] Dec 03 18:50:22 crc kubenswrapper[4787]: E1203 18:50:22.804279 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4006f5-f23a-456a-a376-3ee6fe08a922" containerName="extract-content" Dec 03 18:50:22 crc kubenswrapper[4787]: I1203 18:50:22.804294 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4006f5-f23a-456a-a376-3ee6fe08a922" containerName="extract-content" Dec 03 18:50:22 crc kubenswrapper[4787]: E1203 18:50:22.804331 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4006f5-f23a-456a-a376-3ee6fe08a922" containerName="registry-server" Dec 03 18:50:22 crc kubenswrapper[4787]: I1203 18:50:22.804340 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4006f5-f23a-456a-a376-3ee6fe08a922" containerName="registry-server" Dec 03 18:50:22 crc kubenswrapper[4787]: E1203 18:50:22.804368 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d4006f5-f23a-456a-a376-3ee6fe08a922" containerName="extract-utilities" Dec 03 18:50:22 crc kubenswrapper[4787]: I1203 18:50:22.804379 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d4006f5-f23a-456a-a376-3ee6fe08a922" containerName="extract-utilities" Dec 03 18:50:22 crc kubenswrapper[4787]: I1203 18:50:22.804654 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d4006f5-f23a-456a-a376-3ee6fe08a922" containerName="registry-server" Dec 03 18:50:22 crc kubenswrapper[4787]: I1203 18:50:22.806860 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:22 crc kubenswrapper[4787]: I1203 18:50:22.825512 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-78pj2"] Dec 03 18:50:22 crc kubenswrapper[4787]: I1203 18:50:22.994850 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-utilities\") pod \"certified-operators-78pj2\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:22 crc kubenswrapper[4787]: I1203 18:50:22.994928 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-catalog-content\") pod \"certified-operators-78pj2\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:22 crc kubenswrapper[4787]: I1203 18:50:22.994987 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9zlx\" (UniqueName: \"kubernetes.io/projected/aa481d84-6ada-4778-9464-3516b1f568d5-kube-api-access-t9zlx\") pod \"certified-operators-78pj2\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:23 crc kubenswrapper[4787]: I1203 18:50:23.097169 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-utilities\") pod \"certified-operators-78pj2\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:23 crc kubenswrapper[4787]: I1203 18:50:23.097255 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-catalog-content\") pod \"certified-operators-78pj2\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:23 crc kubenswrapper[4787]: I1203 18:50:23.097309 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9zlx\" (UniqueName: \"kubernetes.io/projected/aa481d84-6ada-4778-9464-3516b1f568d5-kube-api-access-t9zlx\") pod \"certified-operators-78pj2\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:23 crc kubenswrapper[4787]: I1203 18:50:23.098329 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-utilities\") pod \"certified-operators-78pj2\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:23 crc kubenswrapper[4787]: I1203 18:50:23.098733 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-catalog-content\") pod \"certified-operators-78pj2\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:23 crc kubenswrapper[4787]: I1203 18:50:23.118592 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9zlx\" (UniqueName: \"kubernetes.io/projected/aa481d84-6ada-4778-9464-3516b1f568d5-kube-api-access-t9zlx\") pod \"certified-operators-78pj2\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:23 crc kubenswrapper[4787]: I1203 18:50:23.148608 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:23 crc kubenswrapper[4787]: I1203 18:50:23.741993 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-78pj2"] Dec 03 18:50:24 crc kubenswrapper[4787]: I1203 18:50:24.261197 4787 generic.go:334] "Generic (PLEG): container finished" podID="aa481d84-6ada-4778-9464-3516b1f568d5" containerID="ab76e919a3bf05571865803b175108c69dd4eeb27e8e49ae4f17a87aa1103ca8" exitCode=0 Dec 03 18:50:24 crc kubenswrapper[4787]: I1203 18:50:24.261254 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78pj2" event={"ID":"aa481d84-6ada-4778-9464-3516b1f568d5","Type":"ContainerDied","Data":"ab76e919a3bf05571865803b175108c69dd4eeb27e8e49ae4f17a87aa1103ca8"} Dec 03 18:50:24 crc kubenswrapper[4787]: I1203 18:50:24.261571 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78pj2" event={"ID":"aa481d84-6ada-4778-9464-3516b1f568d5","Type":"ContainerStarted","Data":"9448830c591b997d52a3588d1a4930545da6b7b521b0de7b27f38e90f1175bd9"} Dec 03 18:50:25 crc kubenswrapper[4787]: I1203 18:50:25.278645 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78pj2" event={"ID":"aa481d84-6ada-4778-9464-3516b1f568d5","Type":"ContainerStarted","Data":"838911c33c645031538f526136788cbeb04dfc47649166888957080bd8007f65"} Dec 03 18:50:26 crc kubenswrapper[4787]: I1203 18:50:26.298519 4787 generic.go:334] "Generic (PLEG): container finished" podID="aa481d84-6ada-4778-9464-3516b1f568d5" containerID="838911c33c645031538f526136788cbeb04dfc47649166888957080bd8007f65" exitCode=0 Dec 03 18:50:26 crc kubenswrapper[4787]: I1203 18:50:26.298649 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78pj2" event={"ID":"aa481d84-6ada-4778-9464-3516b1f568d5","Type":"ContainerDied","Data":"838911c33c645031538f526136788cbeb04dfc47649166888957080bd8007f65"} Dec 03 18:50:27 crc kubenswrapper[4787]: I1203 18:50:27.311083 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78pj2" event={"ID":"aa481d84-6ada-4778-9464-3516b1f568d5","Type":"ContainerStarted","Data":"1eb24f5ef9fcb9c0471e048aa976ff546daa6e05ac62b333cb9664adae1434e8"} Dec 03 18:50:27 crc kubenswrapper[4787]: I1203 18:50:27.343997 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-78pj2" podStartSLOduration=2.908989033 podStartE2EDuration="5.343978204s" podCreationTimestamp="2025-12-03 18:50:22 +0000 UTC" firstStartedPulling="2025-12-03 18:50:24.264892017 +0000 UTC m=+5841.082362996" lastFinishedPulling="2025-12-03 18:50:26.699881178 +0000 UTC m=+5843.517352167" observedRunningTime="2025-12-03 18:50:27.332282982 +0000 UTC m=+5844.149753951" watchObservedRunningTime="2025-12-03 18:50:27.343978204 +0000 UTC m=+5844.161449153" Dec 03 18:50:33 crc kubenswrapper[4787]: I1203 18:50:33.148876 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:33 crc kubenswrapper[4787]: I1203 18:50:33.149432 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:33 crc kubenswrapper[4787]: I1203 18:50:33.198237 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:33 crc kubenswrapper[4787]: I1203 18:50:33.483345 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:33 crc kubenswrapper[4787]: I1203 18:50:33.537136 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-78pj2"] Dec 03 18:50:35 crc kubenswrapper[4787]: I1203 18:50:35.433211 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-78pj2" podUID="aa481d84-6ada-4778-9464-3516b1f568d5" containerName="registry-server" containerID="cri-o://1eb24f5ef9fcb9c0471e048aa976ff546daa6e05ac62b333cb9664adae1434e8" gracePeriod=2 Dec 03 18:50:36 crc kubenswrapper[4787]: I1203 18:50:36.447286 4787 generic.go:334] "Generic (PLEG): container finished" podID="aa481d84-6ada-4778-9464-3516b1f568d5" containerID="1eb24f5ef9fcb9c0471e048aa976ff546daa6e05ac62b333cb9664adae1434e8" exitCode=0 Dec 03 18:50:36 crc kubenswrapper[4787]: I1203 18:50:36.447372 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78pj2" event={"ID":"aa481d84-6ada-4778-9464-3516b1f568d5","Type":"ContainerDied","Data":"1eb24f5ef9fcb9c0471e048aa976ff546daa6e05ac62b333cb9664adae1434e8"} Dec 03 18:50:36 crc kubenswrapper[4787]: I1203 18:50:36.781263 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:36 crc kubenswrapper[4787]: I1203 18:50:36.941397 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-catalog-content\") pod \"aa481d84-6ada-4778-9464-3516b1f568d5\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " Dec 03 18:50:36 crc kubenswrapper[4787]: I1203 18:50:36.941539 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t9zlx\" (UniqueName: \"kubernetes.io/projected/aa481d84-6ada-4778-9464-3516b1f568d5-kube-api-access-t9zlx\") pod \"aa481d84-6ada-4778-9464-3516b1f568d5\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " Dec 03 18:50:36 crc kubenswrapper[4787]: I1203 18:50:36.941633 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-utilities\") pod \"aa481d84-6ada-4778-9464-3516b1f568d5\" (UID: \"aa481d84-6ada-4778-9464-3516b1f568d5\") " Dec 03 18:50:36 crc kubenswrapper[4787]: I1203 18:50:36.943456 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-utilities" (OuterVolumeSpecName: "utilities") pod "aa481d84-6ada-4778-9464-3516b1f568d5" (UID: "aa481d84-6ada-4778-9464-3516b1f568d5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:50:36 crc kubenswrapper[4787]: I1203 18:50:36.943779 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:50:36 crc kubenswrapper[4787]: I1203 18:50:36.952190 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa481d84-6ada-4778-9464-3516b1f568d5-kube-api-access-t9zlx" (OuterVolumeSpecName: "kube-api-access-t9zlx") pod "aa481d84-6ada-4778-9464-3516b1f568d5" (UID: "aa481d84-6ada-4778-9464-3516b1f568d5"). InnerVolumeSpecName "kube-api-access-t9zlx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:50:36 crc kubenswrapper[4787]: I1203 18:50:36.989502 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "aa481d84-6ada-4778-9464-3516b1f568d5" (UID: "aa481d84-6ada-4778-9464-3516b1f568d5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:50:37 crc kubenswrapper[4787]: I1203 18:50:37.046259 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t9zlx\" (UniqueName: \"kubernetes.io/projected/aa481d84-6ada-4778-9464-3516b1f568d5-kube-api-access-t9zlx\") on node \"crc\" DevicePath \"\"" Dec 03 18:50:37 crc kubenswrapper[4787]: I1203 18:50:37.046296 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/aa481d84-6ada-4778-9464-3516b1f568d5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:50:37 crc kubenswrapper[4787]: I1203 18:50:37.466253 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-78pj2" event={"ID":"aa481d84-6ada-4778-9464-3516b1f568d5","Type":"ContainerDied","Data":"9448830c591b997d52a3588d1a4930545da6b7b521b0de7b27f38e90f1175bd9"} Dec 03 18:50:37 crc kubenswrapper[4787]: I1203 18:50:37.467719 4787 scope.go:117] "RemoveContainer" containerID="1eb24f5ef9fcb9c0471e048aa976ff546daa6e05ac62b333cb9664adae1434e8" Dec 03 18:50:37 crc kubenswrapper[4787]: I1203 18:50:37.466547 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-78pj2" Dec 03 18:50:37 crc kubenswrapper[4787]: I1203 18:50:37.513722 4787 scope.go:117] "RemoveContainer" containerID="838911c33c645031538f526136788cbeb04dfc47649166888957080bd8007f65" Dec 03 18:50:37 crc kubenswrapper[4787]: I1203 18:50:37.533971 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-78pj2"] Dec 03 18:50:37 crc kubenswrapper[4787]: I1203 18:50:37.553191 4787 scope.go:117] "RemoveContainer" containerID="ab76e919a3bf05571865803b175108c69dd4eeb27e8e49ae4f17a87aa1103ca8" Dec 03 18:50:37 crc kubenswrapper[4787]: I1203 18:50:37.588579 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-78pj2"] Dec 03 18:50:37 crc kubenswrapper[4787]: I1203 18:50:37.783761 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa481d84-6ada-4778-9464-3516b1f568d5" path="/var/lib/kubelet/pods/aa481d84-6ada-4778-9464-3516b1f568d5/volumes" Dec 03 18:52:18 crc kubenswrapper[4787]: I1203 18:52:18.989702 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:52:18 crc kubenswrapper[4787]: I1203 18:52:18.990361 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:52:48 crc kubenswrapper[4787]: I1203 18:52:48.989408 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:52:48 crc kubenswrapper[4787]: I1203 18:52:48.990055 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:53:18 crc kubenswrapper[4787]: I1203 18:53:18.990465 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 18:53:18 crc kubenswrapper[4787]: I1203 18:53:18.991480 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 18:53:18 crc kubenswrapper[4787]: I1203 18:53:18.991606 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 18:53:18 crc kubenswrapper[4787]: I1203 18:53:18.993145 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 18:53:18 crc kubenswrapper[4787]: I1203 18:53:18.993241 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" gracePeriod=600 Dec 03 18:53:19 crc kubenswrapper[4787]: E1203 18:53:19.124926 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:53:19 crc kubenswrapper[4787]: I1203 18:53:19.745961 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" exitCode=0 Dec 03 18:53:19 crc kubenswrapper[4787]: I1203 18:53:19.746057 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5"} Dec 03 18:53:19 crc kubenswrapper[4787]: I1203 18:53:19.746104 4787 scope.go:117] "RemoveContainer" containerID="710bb39b7b391e5d3782590ada5ae767a151048d56b184accb8ea5caf20394fa" Dec 03 18:53:19 crc kubenswrapper[4787]: I1203 18:53:19.747075 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:53:19 crc kubenswrapper[4787]: E1203 18:53:19.747435 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:53:33 crc kubenswrapper[4787]: I1203 18:53:33.782886 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:53:33 crc kubenswrapper[4787]: E1203 18:53:33.784003 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:53:48 crc kubenswrapper[4787]: I1203 18:53:48.766460 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:53:48 crc kubenswrapper[4787]: E1203 18:53:48.767809 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:54:02 crc kubenswrapper[4787]: I1203 18:54:02.766282 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:54:02 crc kubenswrapper[4787]: E1203 18:54:02.767257 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:54:13 crc kubenswrapper[4787]: I1203 18:54:13.788046 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:54:13 crc kubenswrapper[4787]: E1203 18:54:13.788842 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:54:26 crc kubenswrapper[4787]: I1203 18:54:26.767133 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:54:26 crc kubenswrapper[4787]: E1203 18:54:26.768386 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:54:38 crc kubenswrapper[4787]: I1203 18:54:38.766271 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:54:38 crc kubenswrapper[4787]: E1203 18:54:38.767419 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.208387 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4jqr4"] Dec 03 18:54:47 crc kubenswrapper[4787]: E1203 18:54:47.210144 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa481d84-6ada-4778-9464-3516b1f568d5" containerName="extract-utilities" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.210162 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa481d84-6ada-4778-9464-3516b1f568d5" containerName="extract-utilities" Dec 03 18:54:47 crc kubenswrapper[4787]: E1203 18:54:47.210170 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa481d84-6ada-4778-9464-3516b1f568d5" containerName="extract-content" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.210176 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa481d84-6ada-4778-9464-3516b1f568d5" containerName="extract-content" Dec 03 18:54:47 crc kubenswrapper[4787]: E1203 18:54:47.210215 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa481d84-6ada-4778-9464-3516b1f568d5" containerName="registry-server" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.210222 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa481d84-6ada-4778-9464-3516b1f568d5" containerName="registry-server" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.210459 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa481d84-6ada-4778-9464-3516b1f568d5" containerName="registry-server" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.212141 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.249329 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4jqr4"] Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.401465 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zsfjq\" (UniqueName: \"kubernetes.io/projected/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-kube-api-access-zsfjq\") pod \"redhat-operators-4jqr4\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.401522 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-catalog-content\") pod \"redhat-operators-4jqr4\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.401557 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-utilities\") pod \"redhat-operators-4jqr4\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.503591 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zsfjq\" (UniqueName: \"kubernetes.io/projected/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-kube-api-access-zsfjq\") pod \"redhat-operators-4jqr4\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.503656 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-catalog-content\") pod \"redhat-operators-4jqr4\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.503683 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-utilities\") pod \"redhat-operators-4jqr4\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.504593 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-utilities\") pod \"redhat-operators-4jqr4\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.504666 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-catalog-content\") pod \"redhat-operators-4jqr4\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.532255 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zsfjq\" (UniqueName: \"kubernetes.io/projected/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-kube-api-access-zsfjq\") pod \"redhat-operators-4jqr4\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:47 crc kubenswrapper[4787]: I1203 18:54:47.543803 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:48 crc kubenswrapper[4787]: I1203 18:54:48.155837 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4jqr4"] Dec 03 18:54:48 crc kubenswrapper[4787]: I1203 18:54:48.920124 4787 generic.go:334] "Generic (PLEG): container finished" podID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerID="d32b1b98492032c3e231b8a123b8093bc1720b07002350185b868c2c5b3f403e" exitCode=0 Dec 03 18:54:48 crc kubenswrapper[4787]: I1203 18:54:48.920183 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jqr4" event={"ID":"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538","Type":"ContainerDied","Data":"d32b1b98492032c3e231b8a123b8093bc1720b07002350185b868c2c5b3f403e"} Dec 03 18:54:48 crc kubenswrapper[4787]: I1203 18:54:48.920388 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jqr4" event={"ID":"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538","Type":"ContainerStarted","Data":"babf595243552797adf4c91cd7563e75b391d1ec089598535a7abc31396ce71c"} Dec 03 18:54:48 crc kubenswrapper[4787]: I1203 18:54:48.922314 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 18:54:49 crc kubenswrapper[4787]: I1203 18:54:49.945292 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jqr4" event={"ID":"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538","Type":"ContainerStarted","Data":"3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d"} Dec 03 18:54:51 crc kubenswrapper[4787]: E1203 18:54:51.184475 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod44b7bc2c_5dd9_4c7c_ba80_cfcaa6ee8538.slice/crio-3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d.scope\": RecentStats: unable to find data in memory cache]" Dec 03 18:54:51 crc kubenswrapper[4787]: I1203 18:54:51.970813 4787 generic.go:334] "Generic (PLEG): container finished" podID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerID="3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d" exitCode=0 Dec 03 18:54:51 crc kubenswrapper[4787]: I1203 18:54:51.970888 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jqr4" event={"ID":"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538","Type":"ContainerDied","Data":"3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d"} Dec 03 18:54:52 crc kubenswrapper[4787]: I1203 18:54:52.766239 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:54:52 crc kubenswrapper[4787]: E1203 18:54:52.766820 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:54:55 crc kubenswrapper[4787]: I1203 18:54:55.016477 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jqr4" event={"ID":"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538","Type":"ContainerStarted","Data":"0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd"} Dec 03 18:54:55 crc kubenswrapper[4787]: I1203 18:54:55.046071 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4jqr4" podStartSLOduration=3.160443846 podStartE2EDuration="8.046054142s" podCreationTimestamp="2025-12-03 18:54:47 +0000 UTC" firstStartedPulling="2025-12-03 18:54:48.922105832 +0000 UTC m=+6105.739576791" lastFinishedPulling="2025-12-03 18:54:53.807716108 +0000 UTC m=+6110.625187087" observedRunningTime="2025-12-03 18:54:55.039433596 +0000 UTC m=+6111.856904595" watchObservedRunningTime="2025-12-03 18:54:55.046054142 +0000 UTC m=+6111.863525101" Dec 03 18:54:57 crc kubenswrapper[4787]: I1203 18:54:57.545388 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:57 crc kubenswrapper[4787]: I1203 18:54:57.545725 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:54:58 crc kubenswrapper[4787]: I1203 18:54:58.621357 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4jqr4" podUID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerName="registry-server" probeResult="failure" output=< Dec 03 18:54:58 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 18:54:58 crc kubenswrapper[4787]: > Dec 03 18:55:04 crc kubenswrapper[4787]: I1203 18:55:04.767161 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:55:04 crc kubenswrapper[4787]: E1203 18:55:04.768051 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:55:07 crc kubenswrapper[4787]: I1203 18:55:07.623649 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:55:07 crc kubenswrapper[4787]: I1203 18:55:07.725701 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:55:07 crc kubenswrapper[4787]: I1203 18:55:07.865551 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4jqr4"] Dec 03 18:55:09 crc kubenswrapper[4787]: I1203 18:55:09.181536 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4jqr4" podUID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerName="registry-server" containerID="cri-o://0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd" gracePeriod=2 Dec 03 18:55:09 crc kubenswrapper[4787]: I1203 18:55:09.784130 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:55:09 crc kubenswrapper[4787]: I1203 18:55:09.838466 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-utilities\") pod \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " Dec 03 18:55:09 crc kubenswrapper[4787]: I1203 18:55:09.838570 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zsfjq\" (UniqueName: \"kubernetes.io/projected/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-kube-api-access-zsfjq\") pod \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " Dec 03 18:55:09 crc kubenswrapper[4787]: I1203 18:55:09.838631 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-catalog-content\") pod \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\" (UID: \"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538\") " Dec 03 18:55:09 crc kubenswrapper[4787]: I1203 18:55:09.841614 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-utilities" (OuterVolumeSpecName: "utilities") pod "44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" (UID: "44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:55:09 crc kubenswrapper[4787]: I1203 18:55:09.863316 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-kube-api-access-zsfjq" (OuterVolumeSpecName: "kube-api-access-zsfjq") pod "44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" (UID: "44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538"). InnerVolumeSpecName "kube-api-access-zsfjq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:55:09 crc kubenswrapper[4787]: I1203 18:55:09.941519 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:55:09 crc kubenswrapper[4787]: I1203 18:55:09.941560 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zsfjq\" (UniqueName: \"kubernetes.io/projected/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-kube-api-access-zsfjq\") on node \"crc\" DevicePath \"\"" Dec 03 18:55:09 crc kubenswrapper[4787]: I1203 18:55:09.975619 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" (UID: "44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.042915 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.192218 4787 generic.go:334] "Generic (PLEG): container finished" podID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerID="0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd" exitCode=0 Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.192292 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4jqr4" Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.192350 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jqr4" event={"ID":"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538","Type":"ContainerDied","Data":"0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd"} Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.192657 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4jqr4" event={"ID":"44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538","Type":"ContainerDied","Data":"babf595243552797adf4c91cd7563e75b391d1ec089598535a7abc31396ce71c"} Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.192687 4787 scope.go:117] "RemoveContainer" containerID="0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd" Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.230878 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4jqr4"] Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.233050 4787 scope.go:117] "RemoveContainer" containerID="3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d" Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.243370 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4jqr4"] Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.259357 4787 scope.go:117] "RemoveContainer" containerID="d32b1b98492032c3e231b8a123b8093bc1720b07002350185b868c2c5b3f403e" Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.325528 4787 scope.go:117] "RemoveContainer" containerID="0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd" Dec 03 18:55:10 crc kubenswrapper[4787]: E1203 18:55:10.326098 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd\": container with ID starting with 0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd not found: ID does not exist" containerID="0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd" Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.326152 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd"} err="failed to get container status \"0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd\": rpc error: code = NotFound desc = could not find container \"0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd\": container with ID starting with 0141daeeaf7793b35ecb8992383fc655d129d2f39120f12f3a21ef7f147c63dd not found: ID does not exist" Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.326176 4787 scope.go:117] "RemoveContainer" containerID="3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d" Dec 03 18:55:10 crc kubenswrapper[4787]: E1203 18:55:10.326674 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d\": container with ID starting with 3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d not found: ID does not exist" containerID="3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d" Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.326723 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d"} err="failed to get container status \"3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d\": rpc error: code = NotFound desc = could not find container \"3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d\": container with ID starting with 3f58efb14c3fd9c31b275649c167836150fc33b9ab57d8f185e8a7336cb95c3d not found: ID does not exist" Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.326752 4787 scope.go:117] "RemoveContainer" containerID="d32b1b98492032c3e231b8a123b8093bc1720b07002350185b868c2c5b3f403e" Dec 03 18:55:10 crc kubenswrapper[4787]: E1203 18:55:10.329134 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d32b1b98492032c3e231b8a123b8093bc1720b07002350185b868c2c5b3f403e\": container with ID starting with d32b1b98492032c3e231b8a123b8093bc1720b07002350185b868c2c5b3f403e not found: ID does not exist" containerID="d32b1b98492032c3e231b8a123b8093bc1720b07002350185b868c2c5b3f403e" Dec 03 18:55:10 crc kubenswrapper[4787]: I1203 18:55:10.329165 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d32b1b98492032c3e231b8a123b8093bc1720b07002350185b868c2c5b3f403e"} err="failed to get container status \"d32b1b98492032c3e231b8a123b8093bc1720b07002350185b868c2c5b3f403e\": rpc error: code = NotFound desc = could not find container \"d32b1b98492032c3e231b8a123b8093bc1720b07002350185b868c2c5b3f403e\": container with ID starting with d32b1b98492032c3e231b8a123b8093bc1720b07002350185b868c2c5b3f403e not found: ID does not exist" Dec 03 18:55:11 crc kubenswrapper[4787]: I1203 18:55:11.780639 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" path="/var/lib/kubelet/pods/44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538/volumes" Dec 03 18:55:12 crc kubenswrapper[4787]: E1203 18:55:12.020776 4787 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.65:58570->38.102.83.65:44689: write tcp 38.102.83.65:58570->38.102.83.65:44689: write: broken pipe Dec 03 18:55:15 crc kubenswrapper[4787]: I1203 18:55:15.766789 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:55:15 crc kubenswrapper[4787]: E1203 18:55:15.767755 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.547407 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 18:55:24 crc kubenswrapper[4787]: E1203 18:55:24.548255 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerName="registry-server" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.548267 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerName="registry-server" Dec 03 18:55:24 crc kubenswrapper[4787]: E1203 18:55:24.548289 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerName="extract-utilities" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.548295 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerName="extract-utilities" Dec 03 18:55:24 crc kubenswrapper[4787]: E1203 18:55:24.548304 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerName="extract-content" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.548310 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerName="extract-content" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.548508 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="44b7bc2c-5dd9-4c7c-ba80-cfcaa6ee8538" containerName="registry-server" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.549245 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.554336 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.554813 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.555058 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mn9nq" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.560839 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.571049 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.626916 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.627069 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.627477 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.627667 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nqwln\" (UniqueName: \"kubernetes.io/projected/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-kube-api-access-nqwln\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.627767 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-config-data\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.627802 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.627821 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.628105 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.628238 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.732947 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.733108 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.733200 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.733317 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nqwln\" (UniqueName: \"kubernetes.io/projected/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-kube-api-access-nqwln\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.733393 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-config-data\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.733432 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.733463 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.733597 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.733653 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.733733 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.734080 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.735205 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.735407 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-config-data\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.735666 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.739371 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ca-certs\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.740167 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config-secret\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.750653 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ssh-key\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.759543 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nqwln\" (UniqueName: \"kubernetes.io/projected/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-kube-api-access-nqwln\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.780447 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tempest-tests-tempest\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " pod="openstack/tempest-tests-tempest" Dec 03 18:55:24 crc kubenswrapper[4787]: I1203 18:55:24.871871 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 18:55:25 crc kubenswrapper[4787]: I1203 18:55:25.380960 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest"] Dec 03 18:55:26 crc kubenswrapper[4787]: I1203 18:55:26.409008 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4c367f96-011e-4a7c-89e6-53c9ed2d1c90","Type":"ContainerStarted","Data":"50730d92b8a739cb20322926f7d6b51e7e525cbf60f5bc6bfa1fbef53c263f0c"} Dec 03 18:55:26 crc kubenswrapper[4787]: I1203 18:55:26.766439 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:55:26 crc kubenswrapper[4787]: E1203 18:55:26.766793 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:55:38 crc kubenswrapper[4787]: I1203 18:55:38.765777 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:55:38 crc kubenswrapper[4787]: E1203 18:55:38.766506 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:55:51 crc kubenswrapper[4787]: I1203 18:55:51.765749 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:55:51 crc kubenswrapper[4787]: E1203 18:55:51.766608 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:56:02 crc kubenswrapper[4787]: E1203 18:56:02.276766 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 03 18:56:02 crc kubenswrapper[4787]: E1203 18:56:02.280455 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nqwln,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest_openstack(4c367f96-011e-4a7c-89e6-53c9ed2d1c90): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 18:56:02 crc kubenswrapper[4787]: E1203 18:56:02.282110 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest" podUID="4c367f96-011e-4a7c-89e6-53c9ed2d1c90" Dec 03 18:56:02 crc kubenswrapper[4787]: E1203 18:56:02.854822 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest" podUID="4c367f96-011e-4a7c-89e6-53c9ed2d1c90" Dec 03 18:56:06 crc kubenswrapper[4787]: I1203 18:56:06.767961 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:56:06 crc kubenswrapper[4787]: E1203 18:56:06.769232 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:56:17 crc kubenswrapper[4787]: I1203 18:56:17.343299 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 03 18:56:19 crc kubenswrapper[4787]: I1203 18:56:19.075060 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4c367f96-011e-4a7c-89e6-53c9ed2d1c90","Type":"ContainerStarted","Data":"883f6485c880565394bf58ea6259a9a888ecffce140118ca938c07df4197e362"} Dec 03 18:56:19 crc kubenswrapper[4787]: I1203 18:56:19.111425 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest" podStartSLOduration=4.166396477 podStartE2EDuration="56.111395153s" podCreationTimestamp="2025-12-03 18:55:23 +0000 UTC" firstStartedPulling="2025-12-03 18:55:25.393692003 +0000 UTC m=+6142.211162962" lastFinishedPulling="2025-12-03 18:56:17.338690649 +0000 UTC m=+6194.156161638" observedRunningTime="2025-12-03 18:56:19.09772479 +0000 UTC m=+6195.915195789" watchObservedRunningTime="2025-12-03 18:56:19.111395153 +0000 UTC m=+6195.928866142" Dec 03 18:56:20 crc kubenswrapper[4787]: I1203 18:56:20.766996 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:56:20 crc kubenswrapper[4787]: E1203 18:56:20.768167 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:56:33 crc kubenswrapper[4787]: I1203 18:56:33.783282 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:56:33 crc kubenswrapper[4787]: E1203 18:56:33.784376 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:56:45 crc kubenswrapper[4787]: I1203 18:56:45.768787 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:56:45 crc kubenswrapper[4787]: E1203 18:56:45.769968 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:57:00 crc kubenswrapper[4787]: I1203 18:57:00.767185 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:57:00 crc kubenswrapper[4787]: E1203 18:57:00.767924 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:57:11 crc kubenswrapper[4787]: I1203 18:57:11.768969 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:57:11 crc kubenswrapper[4787]: E1203 18:57:11.771172 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:57:23 crc kubenswrapper[4787]: I1203 18:57:23.773178 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:57:23 crc kubenswrapper[4787]: E1203 18:57:23.774402 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:57:38 crc kubenswrapper[4787]: I1203 18:57:38.766490 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:57:38 crc kubenswrapper[4787]: E1203 18:57:38.767258 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:57:49 crc kubenswrapper[4787]: I1203 18:57:49.772426 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:57:49 crc kubenswrapper[4787]: E1203 18:57:49.773230 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:58:03 crc kubenswrapper[4787]: I1203 18:58:03.774854 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:58:03 crc kubenswrapper[4787]: E1203 18:58:03.775901 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:58:17 crc kubenswrapper[4787]: I1203 18:58:17.768726 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:58:17 crc kubenswrapper[4787]: E1203 18:58:17.769405 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 18:58:30 crc kubenswrapper[4787]: I1203 18:58:30.766265 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 18:58:31 crc kubenswrapper[4787]: I1203 18:58:31.610363 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"9a7301a11fc6b9d496a1f8decf389ac8d944a317606340a0986691d4155987e3"} Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.678866 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8wnm9"] Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.687698 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.723534 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wnm9"] Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.836961 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28tv8\" (UniqueName: \"kubernetes.io/projected/b9e97d8b-426d-4448-a01a-f210c302f007-kube-api-access-28tv8\") pod \"redhat-marketplace-8wnm9\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.837206 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-utilities\") pod \"redhat-marketplace-8wnm9\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.837387 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-catalog-content\") pod \"redhat-marketplace-8wnm9\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.939806 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-catalog-content\") pod \"redhat-marketplace-8wnm9\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.939968 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28tv8\" (UniqueName: \"kubernetes.io/projected/b9e97d8b-426d-4448-a01a-f210c302f007-kube-api-access-28tv8\") pod \"redhat-marketplace-8wnm9\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.940277 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-utilities\") pod \"redhat-marketplace-8wnm9\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.941975 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-utilities\") pod \"redhat-marketplace-8wnm9\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.942275 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-catalog-content\") pod \"redhat-marketplace-8wnm9\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:13 crc kubenswrapper[4787]: I1203 18:59:13.966163 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28tv8\" (UniqueName: \"kubernetes.io/projected/b9e97d8b-426d-4448-a01a-f210c302f007-kube-api-access-28tv8\") pod \"redhat-marketplace-8wnm9\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:14 crc kubenswrapper[4787]: I1203 18:59:14.012296 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:14 crc kubenswrapper[4787]: I1203 18:59:14.876670 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wnm9"] Dec 03 18:59:14 crc kubenswrapper[4787]: W1203 18:59:14.947278 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9e97d8b_426d_4448_a01a_f210c302f007.slice/crio-68b6a8237ba5ca07a40d27f0111cd5ad1c7af63e081c5186d860e4b100c0dcb2 WatchSource:0}: Error finding container 68b6a8237ba5ca07a40d27f0111cd5ad1c7af63e081c5186d860e4b100c0dcb2: Status 404 returned error can't find the container with id 68b6a8237ba5ca07a40d27f0111cd5ad1c7af63e081c5186d860e4b100c0dcb2 Dec 03 18:59:15 crc kubenswrapper[4787]: I1203 18:59:15.067688 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wnm9" event={"ID":"b9e97d8b-426d-4448-a01a-f210c302f007","Type":"ContainerStarted","Data":"68b6a8237ba5ca07a40d27f0111cd5ad1c7af63e081c5186d860e4b100c0dcb2"} Dec 03 18:59:17 crc kubenswrapper[4787]: I1203 18:59:17.362377 4787 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-6cfhr container/perses-operator namespace/openshift-operators: Liveness probe status=failure output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 18:59:17 crc kubenswrapper[4787]: I1203 18:59:17.364231 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" podUID="3651062d-0769-4da5-bb95-7c5987fc2b7e" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.22:8081/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:59:17 crc kubenswrapper[4787]: I1203 18:59:16.079927 4787 generic.go:334] "Generic (PLEG): container finished" podID="b9e97d8b-426d-4448-a01a-f210c302f007" containerID="2605261ebed3a6af0acb81a5f3a65d5daf55ce8d10ebcab7ea74cf704e9cfdf7" exitCode=0 Dec 03 18:59:17 crc kubenswrapper[4787]: I1203 18:59:17.372162 4787 patch_prober.go:28] interesting pod/perses-operator-5446b9c989-6cfhr container/perses-operator namespace/openshift-operators: Readiness probe status=failure output="Get \"http://10.217.0.22:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 18:59:17 crc kubenswrapper[4787]: I1203 18:59:17.372290 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operators/perses-operator-5446b9c989-6cfhr" podUID="3651062d-0769-4da5-bb95-7c5987fc2b7e" containerName="perses-operator" probeResult="failure" output="Get \"http://10.217.0.22:8081/readyz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 18:59:17 crc kubenswrapper[4787]: I1203 18:59:17.373686 4787 trace.go:236] Trace[135886160]: "Calculate volume metrics of config-volume for pod openshift-dns/dns-default-vjhsg" (03-Dec-2025 18:59:15.916) (total time: 1446ms): Dec 03 18:59:17 crc kubenswrapper[4787]: Trace[135886160]: [1.446055299s] [1.446055299s] END Dec 03 18:59:17 crc kubenswrapper[4787]: E1203 18:59:17.468300 4787 kubelet.go:2526] "Housekeeping took longer than expected" err="housekeeping took too long" expected="1s" actual="1.702s" Dec 03 18:59:17 crc kubenswrapper[4787]: I1203 18:59:17.468413 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wnm9" event={"ID":"b9e97d8b-426d-4448-a01a-f210c302f007","Type":"ContainerDied","Data":"2605261ebed3a6af0acb81a5f3a65d5daf55ce8d10ebcab7ea74cf704e9cfdf7"} Dec 03 18:59:19 crc kubenswrapper[4787]: I1203 18:59:19.387967 4787 generic.go:334] "Generic (PLEG): container finished" podID="b9e97d8b-426d-4448-a01a-f210c302f007" containerID="7ebe0d4322b5c948017d5b212ac0244444fee9f37425f6008a4c3e5ac6a946d6" exitCode=0 Dec 03 18:59:19 crc kubenswrapper[4787]: I1203 18:59:19.388081 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wnm9" event={"ID":"b9e97d8b-426d-4448-a01a-f210c302f007","Type":"ContainerDied","Data":"7ebe0d4322b5c948017d5b212ac0244444fee9f37425f6008a4c3e5ac6a946d6"} Dec 03 18:59:20 crc kubenswrapper[4787]: I1203 18:59:20.400617 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wnm9" event={"ID":"b9e97d8b-426d-4448-a01a-f210c302f007","Type":"ContainerStarted","Data":"1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4"} Dec 03 18:59:20 crc kubenswrapper[4787]: I1203 18:59:20.421677 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8wnm9" podStartSLOduration=3.723392984 podStartE2EDuration="7.421356334s" podCreationTimestamp="2025-12-03 18:59:13 +0000 UTC" firstStartedPulling="2025-12-03 18:59:16.082485936 +0000 UTC m=+6372.899956895" lastFinishedPulling="2025-12-03 18:59:19.780449286 +0000 UTC m=+6376.597920245" observedRunningTime="2025-12-03 18:59:20.420332857 +0000 UTC m=+6377.237803846" watchObservedRunningTime="2025-12-03 18:59:20.421356334 +0000 UTC m=+6377.238827293" Dec 03 18:59:24 crc kubenswrapper[4787]: I1203 18:59:24.012593 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:24 crc kubenswrapper[4787]: I1203 18:59:24.012899 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:24 crc kubenswrapper[4787]: I1203 18:59:24.086480 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:34 crc kubenswrapper[4787]: I1203 18:59:34.077980 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:34 crc kubenswrapper[4787]: I1203 18:59:34.142532 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wnm9"] Dec 03 18:59:34 crc kubenswrapper[4787]: I1203 18:59:34.576923 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8wnm9" podUID="b9e97d8b-426d-4448-a01a-f210c302f007" containerName="registry-server" containerID="cri-o://1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4" gracePeriod=2 Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.464378 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.551200 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-catalog-content\") pod \"b9e97d8b-426d-4448-a01a-f210c302f007\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.551295 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-utilities\") pod \"b9e97d8b-426d-4448-a01a-f210c302f007\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.551348 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-28tv8\" (UniqueName: \"kubernetes.io/projected/b9e97d8b-426d-4448-a01a-f210c302f007-kube-api-access-28tv8\") pod \"b9e97d8b-426d-4448-a01a-f210c302f007\" (UID: \"b9e97d8b-426d-4448-a01a-f210c302f007\") " Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.563892 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-utilities" (OuterVolumeSpecName: "utilities") pod "b9e97d8b-426d-4448-a01a-f210c302f007" (UID: "b9e97d8b-426d-4448-a01a-f210c302f007"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.612381 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b9e97d8b-426d-4448-a01a-f210c302f007" (UID: "b9e97d8b-426d-4448-a01a-f210c302f007"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.656297 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.656331 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b9e97d8b-426d-4448-a01a-f210c302f007-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.660169 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9e97d8b-426d-4448-a01a-f210c302f007-kube-api-access-28tv8" (OuterVolumeSpecName: "kube-api-access-28tv8") pod "b9e97d8b-426d-4448-a01a-f210c302f007" (UID: "b9e97d8b-426d-4448-a01a-f210c302f007"). InnerVolumeSpecName "kube-api-access-28tv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.672214 4787 generic.go:334] "Generic (PLEG): container finished" podID="b9e97d8b-426d-4448-a01a-f210c302f007" containerID="1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4" exitCode=0 Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.672257 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wnm9" event={"ID":"b9e97d8b-426d-4448-a01a-f210c302f007","Type":"ContainerDied","Data":"1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4"} Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.672288 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8wnm9" event={"ID":"b9e97d8b-426d-4448-a01a-f210c302f007","Type":"ContainerDied","Data":"68b6a8237ba5ca07a40d27f0111cd5ad1c7af63e081c5186d860e4b100c0dcb2"} Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.672303 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8wnm9" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.672782 4787 scope.go:117] "RemoveContainer" containerID="1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.722384 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wnm9"] Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.724203 4787 scope.go:117] "RemoveContainer" containerID="7ebe0d4322b5c948017d5b212ac0244444fee9f37425f6008a4c3e5ac6a946d6" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.739378 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8wnm9"] Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.758414 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-28tv8\" (UniqueName: \"kubernetes.io/projected/b9e97d8b-426d-4448-a01a-f210c302f007-kube-api-access-28tv8\") on node \"crc\" DevicePath \"\"" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.759539 4787 scope.go:117] "RemoveContainer" containerID="2605261ebed3a6af0acb81a5f3a65d5daf55ce8d10ebcab7ea74cf704e9cfdf7" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.781534 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9e97d8b-426d-4448-a01a-f210c302f007" path="/var/lib/kubelet/pods/b9e97d8b-426d-4448-a01a-f210c302f007/volumes" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.812233 4787 scope.go:117] "RemoveContainer" containerID="1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4" Dec 03 18:59:35 crc kubenswrapper[4787]: E1203 18:59:35.815853 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4\": container with ID starting with 1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4 not found: ID does not exist" containerID="1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.815913 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4"} err="failed to get container status \"1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4\": rpc error: code = NotFound desc = could not find container \"1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4\": container with ID starting with 1918ac1a7be085b4c14982ceb56915eac01df55932ca21f7a2ea52bd1522c7d4 not found: ID does not exist" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.815943 4787 scope.go:117] "RemoveContainer" containerID="7ebe0d4322b5c948017d5b212ac0244444fee9f37425f6008a4c3e5ac6a946d6" Dec 03 18:59:35 crc kubenswrapper[4787]: E1203 18:59:35.816437 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7ebe0d4322b5c948017d5b212ac0244444fee9f37425f6008a4c3e5ac6a946d6\": container with ID starting with 7ebe0d4322b5c948017d5b212ac0244444fee9f37425f6008a4c3e5ac6a946d6 not found: ID does not exist" containerID="7ebe0d4322b5c948017d5b212ac0244444fee9f37425f6008a4c3e5ac6a946d6" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.816464 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7ebe0d4322b5c948017d5b212ac0244444fee9f37425f6008a4c3e5ac6a946d6"} err="failed to get container status \"7ebe0d4322b5c948017d5b212ac0244444fee9f37425f6008a4c3e5ac6a946d6\": rpc error: code = NotFound desc = could not find container \"7ebe0d4322b5c948017d5b212ac0244444fee9f37425f6008a4c3e5ac6a946d6\": container with ID starting with 7ebe0d4322b5c948017d5b212ac0244444fee9f37425f6008a4c3e5ac6a946d6 not found: ID does not exist" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.816483 4787 scope.go:117] "RemoveContainer" containerID="2605261ebed3a6af0acb81a5f3a65d5daf55ce8d10ebcab7ea74cf704e9cfdf7" Dec 03 18:59:35 crc kubenswrapper[4787]: E1203 18:59:35.816808 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2605261ebed3a6af0acb81a5f3a65d5daf55ce8d10ebcab7ea74cf704e9cfdf7\": container with ID starting with 2605261ebed3a6af0acb81a5f3a65d5daf55ce8d10ebcab7ea74cf704e9cfdf7 not found: ID does not exist" containerID="2605261ebed3a6af0acb81a5f3a65d5daf55ce8d10ebcab7ea74cf704e9cfdf7" Dec 03 18:59:35 crc kubenswrapper[4787]: I1203 18:59:35.816834 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2605261ebed3a6af0acb81a5f3a65d5daf55ce8d10ebcab7ea74cf704e9cfdf7"} err="failed to get container status \"2605261ebed3a6af0acb81a5f3a65d5daf55ce8d10ebcab7ea74cf704e9cfdf7\": rpc error: code = NotFound desc = could not find container \"2605261ebed3a6af0acb81a5f3a65d5daf55ce8d10ebcab7ea74cf704e9cfdf7\": container with ID starting with 2605261ebed3a6af0acb81a5f3a65d5daf55ce8d10ebcab7ea74cf704e9cfdf7 not found: ID does not exist" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.479388 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz"] Dec 03 19:00:00 crc kubenswrapper[4787]: E1203 19:00:00.480586 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9e97d8b-426d-4448-a01a-f210c302f007" containerName="registry-server" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.480719 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9e97d8b-426d-4448-a01a-f210c302f007" containerName="registry-server" Dec 03 19:00:00 crc kubenswrapper[4787]: E1203 19:00:00.480754 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9e97d8b-426d-4448-a01a-f210c302f007" containerName="extract-utilities" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.480761 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9e97d8b-426d-4448-a01a-f210c302f007" containerName="extract-utilities" Dec 03 19:00:00 crc kubenswrapper[4787]: E1203 19:00:00.480785 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9e97d8b-426d-4448-a01a-f210c302f007" containerName="extract-content" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.480793 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9e97d8b-426d-4448-a01a-f210c302f007" containerName="extract-content" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.481204 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9e97d8b-426d-4448-a01a-f210c302f007" containerName="registry-server" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.483049 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.491729 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.493801 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.573707 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz"] Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.680329 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c73281c8-3bf9-4452-a64b-7091dc15a66a-config-volume\") pod \"collect-profiles-29413140-hktmz\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.681101 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c73281c8-3bf9-4452-a64b-7091dc15a66a-secret-volume\") pod \"collect-profiles-29413140-hktmz\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.681363 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6vqn\" (UniqueName: \"kubernetes.io/projected/c73281c8-3bf9-4452-a64b-7091dc15a66a-kube-api-access-g6vqn\") pod \"collect-profiles-29413140-hktmz\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.783763 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c73281c8-3bf9-4452-a64b-7091dc15a66a-config-volume\") pod \"collect-profiles-29413140-hktmz\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.784107 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c73281c8-3bf9-4452-a64b-7091dc15a66a-secret-volume\") pod \"collect-profiles-29413140-hktmz\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.784416 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6vqn\" (UniqueName: \"kubernetes.io/projected/c73281c8-3bf9-4452-a64b-7091dc15a66a-kube-api-access-g6vqn\") pod \"collect-profiles-29413140-hktmz\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.787192 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c73281c8-3bf9-4452-a64b-7091dc15a66a-config-volume\") pod \"collect-profiles-29413140-hktmz\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.799968 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c73281c8-3bf9-4452-a64b-7091dc15a66a-secret-volume\") pod \"collect-profiles-29413140-hktmz\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:00 crc kubenswrapper[4787]: I1203 19:00:00.812162 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6vqn\" (UniqueName: \"kubernetes.io/projected/c73281c8-3bf9-4452-a64b-7091dc15a66a-kube-api-access-g6vqn\") pod \"collect-profiles-29413140-hktmz\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:01 crc kubenswrapper[4787]: I1203 19:00:01.105420 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:01 crc kubenswrapper[4787]: I1203 19:00:01.683609 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz"] Dec 03 19:00:01 crc kubenswrapper[4787]: I1203 19:00:01.996693 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" event={"ID":"c73281c8-3bf9-4452-a64b-7091dc15a66a","Type":"ContainerStarted","Data":"e26a3d3657bdf82ee26940d56ba41ab553500ffe42becde323bc84d8ddfde3a8"} Dec 03 19:00:01 crc kubenswrapper[4787]: I1203 19:00:01.996965 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" event={"ID":"c73281c8-3bf9-4452-a64b-7091dc15a66a","Type":"ContainerStarted","Data":"56cf0254f3a8abcff1061a9098492cab5ad32006a691f92eebe8ea907c75ca71"} Dec 03 19:00:02 crc kubenswrapper[4787]: I1203 19:00:02.020818 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" podStartSLOduration=2.020798297 podStartE2EDuration="2.020798297s" podCreationTimestamp="2025-12-03 19:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 19:00:02.017826448 +0000 UTC m=+6418.835297427" watchObservedRunningTime="2025-12-03 19:00:02.020798297 +0000 UTC m=+6418.838269266" Dec 03 19:00:03 crc kubenswrapper[4787]: I1203 19:00:03.009243 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" event={"ID":"c73281c8-3bf9-4452-a64b-7091dc15a66a","Type":"ContainerDied","Data":"e26a3d3657bdf82ee26940d56ba41ab553500ffe42becde323bc84d8ddfde3a8"} Dec 03 19:00:03 crc kubenswrapper[4787]: I1203 19:00:03.009344 4787 generic.go:334] "Generic (PLEG): container finished" podID="c73281c8-3bf9-4452-a64b-7091dc15a66a" containerID="e26a3d3657bdf82ee26940d56ba41ab553500ffe42becde323bc84d8ddfde3a8" exitCode=0 Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.588870 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.668882 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c73281c8-3bf9-4452-a64b-7091dc15a66a-secret-volume\") pod \"c73281c8-3bf9-4452-a64b-7091dc15a66a\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.669342 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c73281c8-3bf9-4452-a64b-7091dc15a66a-config-volume\") pod \"c73281c8-3bf9-4452-a64b-7091dc15a66a\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.669487 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6vqn\" (UniqueName: \"kubernetes.io/projected/c73281c8-3bf9-4452-a64b-7091dc15a66a-kube-api-access-g6vqn\") pod \"c73281c8-3bf9-4452-a64b-7091dc15a66a\" (UID: \"c73281c8-3bf9-4452-a64b-7091dc15a66a\") " Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.670603 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c73281c8-3bf9-4452-a64b-7091dc15a66a-config-volume" (OuterVolumeSpecName: "config-volume") pod "c73281c8-3bf9-4452-a64b-7091dc15a66a" (UID: "c73281c8-3bf9-4452-a64b-7091dc15a66a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.678594 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c73281c8-3bf9-4452-a64b-7091dc15a66a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c73281c8-3bf9-4452-a64b-7091dc15a66a" (UID: "c73281c8-3bf9-4452-a64b-7091dc15a66a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.678843 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c73281c8-3bf9-4452-a64b-7091dc15a66a-kube-api-access-g6vqn" (OuterVolumeSpecName: "kube-api-access-g6vqn") pod "c73281c8-3bf9-4452-a64b-7091dc15a66a" (UID: "c73281c8-3bf9-4452-a64b-7091dc15a66a"). InnerVolumeSpecName "kube-api-access-g6vqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.761371 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6"] Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.771498 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c73281c8-3bf9-4452-a64b-7091dc15a66a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.771601 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c73281c8-3bf9-4452-a64b-7091dc15a66a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.771660 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6vqn\" (UniqueName: \"kubernetes.io/projected/c73281c8-3bf9-4452-a64b-7091dc15a66a-kube-api-access-g6vqn\") on node \"crc\" DevicePath \"\"" Dec 03 19:00:04 crc kubenswrapper[4787]: I1203 19:00:04.774758 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413095-cc9x6"] Dec 03 19:00:05 crc kubenswrapper[4787]: I1203 19:00:05.031614 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" event={"ID":"c73281c8-3bf9-4452-a64b-7091dc15a66a","Type":"ContainerDied","Data":"56cf0254f3a8abcff1061a9098492cab5ad32006a691f92eebe8ea907c75ca71"} Dec 03 19:00:05 crc kubenswrapper[4787]: I1203 19:00:05.031674 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413140-hktmz" Dec 03 19:00:05 crc kubenswrapper[4787]: I1203 19:00:05.031668 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="56cf0254f3a8abcff1061a9098492cab5ad32006a691f92eebe8ea907c75ca71" Dec 03 19:00:05 crc kubenswrapper[4787]: I1203 19:00:05.780302 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="08a749e7-ff3a-40e3-8a57-418cdda847b7" path="/var/lib/kubelet/pods/08a749e7-ff3a-40e3-8a57-418cdda847b7/volumes" Dec 03 19:00:31 crc kubenswrapper[4787]: I1203 19:00:31.665230 4787 scope.go:117] "RemoveContainer" containerID="60ad081957a6443a259c147724d6ecf88de75ccc4c3400a83ca52686ca64849a" Dec 03 19:00:48 crc kubenswrapper[4787]: I1203 19:00:48.991161 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:00:48 crc kubenswrapper[4787]: I1203 19:00:48.992284 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.314317 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-ktqw2"] Dec 03 19:00:58 crc kubenswrapper[4787]: E1203 19:00:58.315758 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c73281c8-3bf9-4452-a64b-7091dc15a66a" containerName="collect-profiles" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.315779 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c73281c8-3bf9-4452-a64b-7091dc15a66a" containerName="collect-profiles" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.316199 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c73281c8-3bf9-4452-a64b-7091dc15a66a" containerName="collect-profiles" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.318725 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.332872 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ktqw2"] Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.400065 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-catalog-content\") pod \"certified-operators-ktqw2\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.400251 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-utilities\") pod \"certified-operators-ktqw2\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.400356 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncwdz\" (UniqueName: \"kubernetes.io/projected/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-kube-api-access-ncwdz\") pod \"certified-operators-ktqw2\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.502977 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-utilities\") pod \"certified-operators-ktqw2\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.503125 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncwdz\" (UniqueName: \"kubernetes.io/projected/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-kube-api-access-ncwdz\") pod \"certified-operators-ktqw2\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.503256 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-catalog-content\") pod \"certified-operators-ktqw2\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.503506 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-utilities\") pod \"certified-operators-ktqw2\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.504074 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-catalog-content\") pod \"certified-operators-ktqw2\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.526226 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncwdz\" (UniqueName: \"kubernetes.io/projected/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-kube-api-access-ncwdz\") pod \"certified-operators-ktqw2\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:00:58 crc kubenswrapper[4787]: I1203 19:00:58.642866 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:00:59 crc kubenswrapper[4787]: I1203 19:00:59.150313 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-ktqw2"] Dec 03 19:00:59 crc kubenswrapper[4787]: W1203 19:00:59.155093 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fa3425b_feb1_4f4c_9cda_10dd9a58357a.slice/crio-5fad7817ac9a64ef64fd0a8d8b0857b344d22f1ac15b100c2733763bcb9ef3b7 WatchSource:0}: Error finding container 5fad7817ac9a64ef64fd0a8d8b0857b344d22f1ac15b100c2733763bcb9ef3b7: Status 404 returned error can't find the container with id 5fad7817ac9a64ef64fd0a8d8b0857b344d22f1ac15b100c2733763bcb9ef3b7 Dec 03 19:00:59 crc kubenswrapper[4787]: I1203 19:00:59.677129 4787 generic.go:334] "Generic (PLEG): container finished" podID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" containerID="1ac7b88d206f401af8854587aae1b9c1f930f6a3a40afb1ff57ebf270fcb7f3b" exitCode=0 Dec 03 19:00:59 crc kubenswrapper[4787]: I1203 19:00:59.677623 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktqw2" event={"ID":"5fa3425b-feb1-4f4c-9cda-10dd9a58357a","Type":"ContainerDied","Data":"1ac7b88d206f401af8854587aae1b9c1f930f6a3a40afb1ff57ebf270fcb7f3b"} Dec 03 19:00:59 crc kubenswrapper[4787]: I1203 19:00:59.678353 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktqw2" event={"ID":"5fa3425b-feb1-4f4c-9cda-10dd9a58357a","Type":"ContainerStarted","Data":"5fad7817ac9a64ef64fd0a8d8b0857b344d22f1ac15b100c2733763bcb9ef3b7"} Dec 03 19:00:59 crc kubenswrapper[4787]: I1203 19:00:59.681619 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.185249 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29413141-r9f29"] Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.187477 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.195721 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413141-r9f29"] Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.353295 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-config-data\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.353545 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-combined-ca-bundle\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.353648 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p74vw\" (UniqueName: \"kubernetes.io/projected/10d9778c-c2bf-40cf-8934-d584cbc0bad7-kube-api-access-p74vw\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.355532 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-fernet-keys\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.457524 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-config-data\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.457608 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-combined-ca-bundle\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.457646 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p74vw\" (UniqueName: \"kubernetes.io/projected/10d9778c-c2bf-40cf-8934-d584cbc0bad7-kube-api-access-p74vw\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.457701 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-fernet-keys\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.465788 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-fernet-keys\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.465828 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-combined-ca-bundle\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.466229 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-config-data\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.479784 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p74vw\" (UniqueName: \"kubernetes.io/projected/10d9778c-c2bf-40cf-8934-d584cbc0bad7-kube-api-access-p74vw\") pod \"keystone-cron-29413141-r9f29\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:00 crc kubenswrapper[4787]: I1203 19:01:00.517618 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:01 crc kubenswrapper[4787]: I1203 19:01:01.084988 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29413141-r9f29"] Dec 03 19:01:01 crc kubenswrapper[4787]: I1203 19:01:01.704665 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413141-r9f29" event={"ID":"10d9778c-c2bf-40cf-8934-d584cbc0bad7","Type":"ContainerStarted","Data":"cb05bbf18906369c0bb945dca725d4798ac526ff142a344de11bd98013364c68"} Dec 03 19:01:01 crc kubenswrapper[4787]: I1203 19:01:01.705155 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413141-r9f29" event={"ID":"10d9778c-c2bf-40cf-8934-d584cbc0bad7","Type":"ContainerStarted","Data":"b617a35b44eee9e9a901caa8321dcb9013c97ad8eb6f1df4014f1e933453e1e5"} Dec 03 19:01:01 crc kubenswrapper[4787]: I1203 19:01:01.709811 4787 generic.go:334] "Generic (PLEG): container finished" podID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" containerID="2199d55d92b4520740637507fd5835c4ce38b88ffa44f590d8cc0bcab686eb56" exitCode=0 Dec 03 19:01:01 crc kubenswrapper[4787]: I1203 19:01:01.709860 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktqw2" event={"ID":"5fa3425b-feb1-4f4c-9cda-10dd9a58357a","Type":"ContainerDied","Data":"2199d55d92b4520740637507fd5835c4ce38b88ffa44f590d8cc0bcab686eb56"} Dec 03 19:01:01 crc kubenswrapper[4787]: I1203 19:01:01.733912 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29413141-r9f29" podStartSLOduration=1.733869977 podStartE2EDuration="1.733869977s" podCreationTimestamp="2025-12-03 19:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 19:01:01.721497378 +0000 UTC m=+6478.538968337" watchObservedRunningTime="2025-12-03 19:01:01.733869977 +0000 UTC m=+6478.551340936" Dec 03 19:01:02 crc kubenswrapper[4787]: I1203 19:01:02.725320 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktqw2" event={"ID":"5fa3425b-feb1-4f4c-9cda-10dd9a58357a","Type":"ContainerStarted","Data":"69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135"} Dec 03 19:01:02 crc kubenswrapper[4787]: I1203 19:01:02.748240 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-ktqw2" podStartSLOduration=2.29094454 podStartE2EDuration="4.748220873s" podCreationTimestamp="2025-12-03 19:00:58 +0000 UTC" firstStartedPulling="2025-12-03 19:00:59.680391127 +0000 UTC m=+6476.497862096" lastFinishedPulling="2025-12-03 19:01:02.13766747 +0000 UTC m=+6478.955138429" observedRunningTime="2025-12-03 19:01:02.742758078 +0000 UTC m=+6479.560229037" watchObservedRunningTime="2025-12-03 19:01:02.748220873 +0000 UTC m=+6479.565691832" Dec 03 19:01:04 crc kubenswrapper[4787]: I1203 19:01:04.794939 4787 generic.go:334] "Generic (PLEG): container finished" podID="10d9778c-c2bf-40cf-8934-d584cbc0bad7" containerID="cb05bbf18906369c0bb945dca725d4798ac526ff142a344de11bd98013364c68" exitCode=0 Dec 03 19:01:04 crc kubenswrapper[4787]: I1203 19:01:04.795256 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413141-r9f29" event={"ID":"10d9778c-c2bf-40cf-8934-d584cbc0bad7","Type":"ContainerDied","Data":"cb05bbf18906369c0bb945dca725d4798ac526ff142a344de11bd98013364c68"} Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.292802 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.420346 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p74vw\" (UniqueName: \"kubernetes.io/projected/10d9778c-c2bf-40cf-8934-d584cbc0bad7-kube-api-access-p74vw\") pod \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.420515 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-combined-ca-bundle\") pod \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.420609 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-fernet-keys\") pod \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.420759 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-config-data\") pod \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\" (UID: \"10d9778c-c2bf-40cf-8934-d584cbc0bad7\") " Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.426488 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "10d9778c-c2bf-40cf-8934-d584cbc0bad7" (UID: "10d9778c-c2bf-40cf-8934-d584cbc0bad7"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.427451 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/10d9778c-c2bf-40cf-8934-d584cbc0bad7-kube-api-access-p74vw" (OuterVolumeSpecName: "kube-api-access-p74vw") pod "10d9778c-c2bf-40cf-8934-d584cbc0bad7" (UID: "10d9778c-c2bf-40cf-8934-d584cbc0bad7"). InnerVolumeSpecName "kube-api-access-p74vw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.464770 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "10d9778c-c2bf-40cf-8934-d584cbc0bad7" (UID: "10d9778c-c2bf-40cf-8934-d584cbc0bad7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.484925 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-config-data" (OuterVolumeSpecName: "config-data") pod "10d9778c-c2bf-40cf-8934-d584cbc0bad7" (UID: "10d9778c-c2bf-40cf-8934-d584cbc0bad7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.523791 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.523849 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p74vw\" (UniqueName: \"kubernetes.io/projected/10d9778c-c2bf-40cf-8934-d584cbc0bad7-kube-api-access-p74vw\") on node \"crc\" DevicePath \"\"" Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.523871 4787 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.523892 4787 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/10d9778c-c2bf-40cf-8934-d584cbc0bad7-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.821430 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29413141-r9f29" event={"ID":"10d9778c-c2bf-40cf-8934-d584cbc0bad7","Type":"ContainerDied","Data":"b617a35b44eee9e9a901caa8321dcb9013c97ad8eb6f1df4014f1e933453e1e5"} Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.821837 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b617a35b44eee9e9a901caa8321dcb9013c97ad8eb6f1df4014f1e933453e1e5" Dec 03 19:01:06 crc kubenswrapper[4787]: I1203 19:01:06.821581 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29413141-r9f29" Dec 03 19:01:08 crc kubenswrapper[4787]: I1203 19:01:08.643522 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:01:08 crc kubenswrapper[4787]: I1203 19:01:08.644008 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:01:08 crc kubenswrapper[4787]: I1203 19:01:08.736307 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:01:08 crc kubenswrapper[4787]: I1203 19:01:08.918236 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:01:08 crc kubenswrapper[4787]: I1203 19:01:08.985731 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ktqw2"] Dec 03 19:01:10 crc kubenswrapper[4787]: I1203 19:01:10.868624 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-ktqw2" podUID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" containerName="registry-server" containerID="cri-o://69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135" gracePeriod=2 Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.480494 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.641327 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-catalog-content\") pod \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.641422 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-utilities\") pod \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.641472 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncwdz\" (UniqueName: \"kubernetes.io/projected/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-kube-api-access-ncwdz\") pod \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\" (UID: \"5fa3425b-feb1-4f4c-9cda-10dd9a58357a\") " Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.642253 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-utilities" (OuterVolumeSpecName: "utilities") pod "5fa3425b-feb1-4f4c-9cda-10dd9a58357a" (UID: "5fa3425b-feb1-4f4c-9cda-10dd9a58357a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.661803 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-kube-api-access-ncwdz" (OuterVolumeSpecName: "kube-api-access-ncwdz") pod "5fa3425b-feb1-4f4c-9cda-10dd9a58357a" (UID: "5fa3425b-feb1-4f4c-9cda-10dd9a58357a"). InnerVolumeSpecName "kube-api-access-ncwdz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.685504 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5fa3425b-feb1-4f4c-9cda-10dd9a58357a" (UID: "5fa3425b-feb1-4f4c-9cda-10dd9a58357a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.749416 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.750102 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.750326 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncwdz\" (UniqueName: \"kubernetes.io/projected/5fa3425b-feb1-4f4c-9cda-10dd9a58357a-kube-api-access-ncwdz\") on node \"crc\" DevicePath \"\"" Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.890296 4787 generic.go:334] "Generic (PLEG): container finished" podID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" containerID="69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135" exitCode=0 Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.890362 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktqw2" event={"ID":"5fa3425b-feb1-4f4c-9cda-10dd9a58357a","Type":"ContainerDied","Data":"69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135"} Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.890403 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-ktqw2" event={"ID":"5fa3425b-feb1-4f4c-9cda-10dd9a58357a","Type":"ContainerDied","Data":"5fad7817ac9a64ef64fd0a8d8b0857b344d22f1ac15b100c2733763bcb9ef3b7"} Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.890438 4787 scope.go:117] "RemoveContainer" containerID="69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135" Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.890974 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-ktqw2" Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.926759 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-ktqw2"] Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.934769 4787 scope.go:117] "RemoveContainer" containerID="2199d55d92b4520740637507fd5835c4ce38b88ffa44f590d8cc0bcab686eb56" Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.979081 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-ktqw2"] Dec 03 19:01:11 crc kubenswrapper[4787]: I1203 19:01:11.981178 4787 scope.go:117] "RemoveContainer" containerID="1ac7b88d206f401af8854587aae1b9c1f930f6a3a40afb1ff57ebf270fcb7f3b" Dec 03 19:01:12 crc kubenswrapper[4787]: I1203 19:01:12.030117 4787 scope.go:117] "RemoveContainer" containerID="69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135" Dec 03 19:01:12 crc kubenswrapper[4787]: E1203 19:01:12.031953 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135\": container with ID starting with 69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135 not found: ID does not exist" containerID="69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135" Dec 03 19:01:12 crc kubenswrapper[4787]: I1203 19:01:12.032003 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135"} err="failed to get container status \"69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135\": rpc error: code = NotFound desc = could not find container \"69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135\": container with ID starting with 69843612002fb7a0519cc0156dfa7d6cc3ac8deffca90c903ea3ccb44f8e3135 not found: ID does not exist" Dec 03 19:01:12 crc kubenswrapper[4787]: I1203 19:01:12.032058 4787 scope.go:117] "RemoveContainer" containerID="2199d55d92b4520740637507fd5835c4ce38b88ffa44f590d8cc0bcab686eb56" Dec 03 19:01:12 crc kubenswrapper[4787]: E1203 19:01:12.033336 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2199d55d92b4520740637507fd5835c4ce38b88ffa44f590d8cc0bcab686eb56\": container with ID starting with 2199d55d92b4520740637507fd5835c4ce38b88ffa44f590d8cc0bcab686eb56 not found: ID does not exist" containerID="2199d55d92b4520740637507fd5835c4ce38b88ffa44f590d8cc0bcab686eb56" Dec 03 19:01:12 crc kubenswrapper[4787]: I1203 19:01:12.033642 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2199d55d92b4520740637507fd5835c4ce38b88ffa44f590d8cc0bcab686eb56"} err="failed to get container status \"2199d55d92b4520740637507fd5835c4ce38b88ffa44f590d8cc0bcab686eb56\": rpc error: code = NotFound desc = could not find container \"2199d55d92b4520740637507fd5835c4ce38b88ffa44f590d8cc0bcab686eb56\": container with ID starting with 2199d55d92b4520740637507fd5835c4ce38b88ffa44f590d8cc0bcab686eb56 not found: ID does not exist" Dec 03 19:01:12 crc kubenswrapper[4787]: I1203 19:01:12.033917 4787 scope.go:117] "RemoveContainer" containerID="1ac7b88d206f401af8854587aae1b9c1f930f6a3a40afb1ff57ebf270fcb7f3b" Dec 03 19:01:12 crc kubenswrapper[4787]: E1203 19:01:12.034669 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ac7b88d206f401af8854587aae1b9c1f930f6a3a40afb1ff57ebf270fcb7f3b\": container with ID starting with 1ac7b88d206f401af8854587aae1b9c1f930f6a3a40afb1ff57ebf270fcb7f3b not found: ID does not exist" containerID="1ac7b88d206f401af8854587aae1b9c1f930f6a3a40afb1ff57ebf270fcb7f3b" Dec 03 19:01:12 crc kubenswrapper[4787]: I1203 19:01:12.034735 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ac7b88d206f401af8854587aae1b9c1f930f6a3a40afb1ff57ebf270fcb7f3b"} err="failed to get container status \"1ac7b88d206f401af8854587aae1b9c1f930f6a3a40afb1ff57ebf270fcb7f3b\": rpc error: code = NotFound desc = could not find container \"1ac7b88d206f401af8854587aae1b9c1f930f6a3a40afb1ff57ebf270fcb7f3b\": container with ID starting with 1ac7b88d206f401af8854587aae1b9c1f930f6a3a40afb1ff57ebf270fcb7f3b not found: ID does not exist" Dec 03 19:01:13 crc kubenswrapper[4787]: I1203 19:01:13.794679 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" path="/var/lib/kubelet/pods/5fa3425b-feb1-4f4c-9cda-10dd9a58357a/volumes" Dec 03 19:01:18 crc kubenswrapper[4787]: I1203 19:01:18.996219 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:01:18 crc kubenswrapper[4787]: I1203 19:01:18.996970 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:01:48 crc kubenswrapper[4787]: I1203 19:01:48.989590 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:01:48 crc kubenswrapper[4787]: I1203 19:01:48.990280 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:01:48 crc kubenswrapper[4787]: I1203 19:01:48.990354 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 19:01:48 crc kubenswrapper[4787]: I1203 19:01:48.991533 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"9a7301a11fc6b9d496a1f8decf389ac8d944a317606340a0986691d4155987e3"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:01:48 crc kubenswrapper[4787]: I1203 19:01:48.991654 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://9a7301a11fc6b9d496a1f8decf389ac8d944a317606340a0986691d4155987e3" gracePeriod=600 Dec 03 19:01:49 crc kubenswrapper[4787]: I1203 19:01:49.337481 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="9a7301a11fc6b9d496a1f8decf389ac8d944a317606340a0986691d4155987e3" exitCode=0 Dec 03 19:01:49 crc kubenswrapper[4787]: I1203 19:01:49.337589 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"9a7301a11fc6b9d496a1f8decf389ac8d944a317606340a0986691d4155987e3"} Dec 03 19:01:49 crc kubenswrapper[4787]: I1203 19:01:49.337859 4787 scope.go:117] "RemoveContainer" containerID="979a62f32924f216343c05f7c393b6781d744364a45669e59a85cd88410ebea5" Dec 03 19:01:50 crc kubenswrapper[4787]: I1203 19:01:50.351331 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28"} Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.365466 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-gfs8n"] Dec 03 19:03:11 crc kubenswrapper[4787]: E1203 19:03:11.366920 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="10d9778c-c2bf-40cf-8934-d584cbc0bad7" containerName="keystone-cron" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.366946 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="10d9778c-c2bf-40cf-8934-d584cbc0bad7" containerName="keystone-cron" Dec 03 19:03:11 crc kubenswrapper[4787]: E1203 19:03:11.366976 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" containerName="extract-content" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.366989 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" containerName="extract-content" Dec 03 19:03:11 crc kubenswrapper[4787]: E1203 19:03:11.367012 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" containerName="registry-server" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.367061 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" containerName="registry-server" Dec 03 19:03:11 crc kubenswrapper[4787]: E1203 19:03:11.367153 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" containerName="extract-utilities" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.367167 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" containerName="extract-utilities" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.367548 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="10d9778c-c2bf-40cf-8934-d584cbc0bad7" containerName="keystone-cron" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.367608 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="5fa3425b-feb1-4f4c-9cda-10dd9a58357a" containerName="registry-server" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.370833 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.381155 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gfs8n"] Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.423571 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-catalog-content\") pod \"community-operators-gfs8n\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.423638 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-utilities\") pod \"community-operators-gfs8n\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.423928 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2ddjw\" (UniqueName: \"kubernetes.io/projected/861cf19a-5812-4179-9eb5-436dfbc561a8-kube-api-access-2ddjw\") pod \"community-operators-gfs8n\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.527368 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-catalog-content\") pod \"community-operators-gfs8n\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.527439 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-utilities\") pod \"community-operators-gfs8n\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.527557 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2ddjw\" (UniqueName: \"kubernetes.io/projected/861cf19a-5812-4179-9eb5-436dfbc561a8-kube-api-access-2ddjw\") pod \"community-operators-gfs8n\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.528203 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-utilities\") pod \"community-operators-gfs8n\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.528335 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-catalog-content\") pod \"community-operators-gfs8n\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.548786 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2ddjw\" (UniqueName: \"kubernetes.io/projected/861cf19a-5812-4179-9eb5-436dfbc561a8-kube-api-access-2ddjw\") pod \"community-operators-gfs8n\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:11 crc kubenswrapper[4787]: I1203 19:03:11.703947 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:12 crc kubenswrapper[4787]: I1203 19:03:12.205577 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-gfs8n"] Dec 03 19:03:12 crc kubenswrapper[4787]: I1203 19:03:12.343158 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfs8n" event={"ID":"861cf19a-5812-4179-9eb5-436dfbc561a8","Type":"ContainerStarted","Data":"cf69435ec749b3b00f501086b724cdb534ffcb58dae62c5aed4fa6b8aee5b34c"} Dec 03 19:03:13 crc kubenswrapper[4787]: I1203 19:03:13.355224 4787 generic.go:334] "Generic (PLEG): container finished" podID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerID="4282f037bc33c159598bc1cb1fd476d310a75d0a0698f12b3cdfcac01920be2a" exitCode=0 Dec 03 19:03:13 crc kubenswrapper[4787]: I1203 19:03:13.355303 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfs8n" event={"ID":"861cf19a-5812-4179-9eb5-436dfbc561a8","Type":"ContainerDied","Data":"4282f037bc33c159598bc1cb1fd476d310a75d0a0698f12b3cdfcac01920be2a"} Dec 03 19:03:15 crc kubenswrapper[4787]: I1203 19:03:15.381678 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfs8n" event={"ID":"861cf19a-5812-4179-9eb5-436dfbc561a8","Type":"ContainerStarted","Data":"fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e"} Dec 03 19:03:16 crc kubenswrapper[4787]: I1203 19:03:16.396561 4787 generic.go:334] "Generic (PLEG): container finished" podID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerID="fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e" exitCode=0 Dec 03 19:03:16 crc kubenswrapper[4787]: I1203 19:03:16.396864 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfs8n" event={"ID":"861cf19a-5812-4179-9eb5-436dfbc561a8","Type":"ContainerDied","Data":"fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e"} Dec 03 19:03:17 crc kubenswrapper[4787]: I1203 19:03:17.408705 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfs8n" event={"ID":"861cf19a-5812-4179-9eb5-436dfbc561a8","Type":"ContainerStarted","Data":"4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e"} Dec 03 19:03:17 crc kubenswrapper[4787]: I1203 19:03:17.427741 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-gfs8n" podStartSLOduration=2.997039653 podStartE2EDuration="6.427725734s" podCreationTimestamp="2025-12-03 19:03:11 +0000 UTC" firstStartedPulling="2025-12-03 19:03:13.3575331 +0000 UTC m=+6610.175004059" lastFinishedPulling="2025-12-03 19:03:16.788219181 +0000 UTC m=+6613.605690140" observedRunningTime="2025-12-03 19:03:17.426377708 +0000 UTC m=+6614.243848677" watchObservedRunningTime="2025-12-03 19:03:17.427725734 +0000 UTC m=+6614.245196693" Dec 03 19:03:21 crc kubenswrapper[4787]: I1203 19:03:21.704877 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:21 crc kubenswrapper[4787]: I1203 19:03:21.707788 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:22 crc kubenswrapper[4787]: I1203 19:03:22.805280 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-gfs8n" podUID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerName="registry-server" probeResult="failure" output=< Dec 03 19:03:22 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 19:03:22 crc kubenswrapper[4787]: > Dec 03 19:03:31 crc kubenswrapper[4787]: I1203 19:03:31.782111 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:31 crc kubenswrapper[4787]: I1203 19:03:31.834336 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:32 crc kubenswrapper[4787]: I1203 19:03:32.028678 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gfs8n"] Dec 03 19:03:33 crc kubenswrapper[4787]: I1203 19:03:33.587091 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-gfs8n" podUID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerName="registry-server" containerID="cri-o://4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e" gracePeriod=2 Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.246915 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.293193 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2ddjw\" (UniqueName: \"kubernetes.io/projected/861cf19a-5812-4179-9eb5-436dfbc561a8-kube-api-access-2ddjw\") pod \"861cf19a-5812-4179-9eb5-436dfbc561a8\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.293481 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-utilities\") pod \"861cf19a-5812-4179-9eb5-436dfbc561a8\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.293701 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-catalog-content\") pod \"861cf19a-5812-4179-9eb5-436dfbc561a8\" (UID: \"861cf19a-5812-4179-9eb5-436dfbc561a8\") " Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.294853 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-utilities" (OuterVolumeSpecName: "utilities") pod "861cf19a-5812-4179-9eb5-436dfbc561a8" (UID: "861cf19a-5812-4179-9eb5-436dfbc561a8"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.304414 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/861cf19a-5812-4179-9eb5-436dfbc561a8-kube-api-access-2ddjw" (OuterVolumeSpecName: "kube-api-access-2ddjw") pod "861cf19a-5812-4179-9eb5-436dfbc561a8" (UID: "861cf19a-5812-4179-9eb5-436dfbc561a8"). InnerVolumeSpecName "kube-api-access-2ddjw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.381606 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "861cf19a-5812-4179-9eb5-436dfbc561a8" (UID: "861cf19a-5812-4179-9eb5-436dfbc561a8"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.396105 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2ddjw\" (UniqueName: \"kubernetes.io/projected/861cf19a-5812-4179-9eb5-436dfbc561a8-kube-api-access-2ddjw\") on node \"crc\" DevicePath \"\"" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.396140 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.396152 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/861cf19a-5812-4179-9eb5-436dfbc561a8-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.614296 4787 generic.go:334] "Generic (PLEG): container finished" podID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerID="4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e" exitCode=0 Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.614404 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfs8n" event={"ID":"861cf19a-5812-4179-9eb5-436dfbc561a8","Type":"ContainerDied","Data":"4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e"} Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.614443 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-gfs8n" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.614650 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-gfs8n" event={"ID":"861cf19a-5812-4179-9eb5-436dfbc561a8","Type":"ContainerDied","Data":"cf69435ec749b3b00f501086b724cdb534ffcb58dae62c5aed4fa6b8aee5b34c"} Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.614682 4787 scope.go:117] "RemoveContainer" containerID="4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.654283 4787 scope.go:117] "RemoveContainer" containerID="fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.662959 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-gfs8n"] Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.675832 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-gfs8n"] Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.701579 4787 scope.go:117] "RemoveContainer" containerID="4282f037bc33c159598bc1cb1fd476d310a75d0a0698f12b3cdfcac01920be2a" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.761601 4787 scope.go:117] "RemoveContainer" containerID="4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e" Dec 03 19:03:34 crc kubenswrapper[4787]: E1203 19:03:34.762076 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e\": container with ID starting with 4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e not found: ID does not exist" containerID="4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.762115 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e"} err="failed to get container status \"4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e\": rpc error: code = NotFound desc = could not find container \"4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e\": container with ID starting with 4fedbec9db1501ddd6330539c559793aeef7ab0982d0af6484fa80959381d76e not found: ID does not exist" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.762139 4787 scope.go:117] "RemoveContainer" containerID="fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e" Dec 03 19:03:34 crc kubenswrapper[4787]: E1203 19:03:34.762480 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e\": container with ID starting with fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e not found: ID does not exist" containerID="fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.762513 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e"} err="failed to get container status \"fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e\": rpc error: code = NotFound desc = could not find container \"fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e\": container with ID starting with fb4c8d19738323c08aa1060e3509fbcaabdebd4148bd005b443cf7a52b7d3b4e not found: ID does not exist" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.762531 4787 scope.go:117] "RemoveContainer" containerID="4282f037bc33c159598bc1cb1fd476d310a75d0a0698f12b3cdfcac01920be2a" Dec 03 19:03:34 crc kubenswrapper[4787]: E1203 19:03:34.762900 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4282f037bc33c159598bc1cb1fd476d310a75d0a0698f12b3cdfcac01920be2a\": container with ID starting with 4282f037bc33c159598bc1cb1fd476d310a75d0a0698f12b3cdfcac01920be2a not found: ID does not exist" containerID="4282f037bc33c159598bc1cb1fd476d310a75d0a0698f12b3cdfcac01920be2a" Dec 03 19:03:34 crc kubenswrapper[4787]: I1203 19:03:34.762926 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4282f037bc33c159598bc1cb1fd476d310a75d0a0698f12b3cdfcac01920be2a"} err="failed to get container status \"4282f037bc33c159598bc1cb1fd476d310a75d0a0698f12b3cdfcac01920be2a\": rpc error: code = NotFound desc = could not find container \"4282f037bc33c159598bc1cb1fd476d310a75d0a0698f12b3cdfcac01920be2a\": container with ID starting with 4282f037bc33c159598bc1cb1fd476d310a75d0a0698f12b3cdfcac01920be2a not found: ID does not exist" Dec 03 19:03:35 crc kubenswrapper[4787]: I1203 19:03:35.778969 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="861cf19a-5812-4179-9eb5-436dfbc561a8" path="/var/lib/kubelet/pods/861cf19a-5812-4179-9eb5-436dfbc561a8/volumes" Dec 03 19:04:18 crc kubenswrapper[4787]: I1203 19:04:18.989583 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:04:18 crc kubenswrapper[4787]: I1203 19:04:18.990050 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:04:48 crc kubenswrapper[4787]: I1203 19:04:48.990069 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:04:48 crc kubenswrapper[4787]: I1203 19:04:48.990696 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:05:18 crc kubenswrapper[4787]: I1203 19:05:18.990559 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:05:18 crc kubenswrapper[4787]: I1203 19:05:18.991165 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:05:18 crc kubenswrapper[4787]: I1203 19:05:18.991226 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 19:05:18 crc kubenswrapper[4787]: I1203 19:05:18.992258 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:05:18 crc kubenswrapper[4787]: I1203 19:05:18.992326 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" gracePeriod=600 Dec 03 19:05:19 crc kubenswrapper[4787]: E1203 19:05:19.116990 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:05:20 crc kubenswrapper[4787]: I1203 19:05:20.070777 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" exitCode=0 Dec 03 19:05:20 crc kubenswrapper[4787]: I1203 19:05:20.070878 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28"} Dec 03 19:05:20 crc kubenswrapper[4787]: I1203 19:05:20.071126 4787 scope.go:117] "RemoveContainer" containerID="9a7301a11fc6b9d496a1f8decf389ac8d944a317606340a0986691d4155987e3" Dec 03 19:05:20 crc kubenswrapper[4787]: I1203 19:05:20.072074 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:05:20 crc kubenswrapper[4787]: E1203 19:05:20.072438 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:05:31 crc kubenswrapper[4787]: I1203 19:05:31.766736 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:05:31 crc kubenswrapper[4787]: E1203 19:05:31.767754 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.686420 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-nkm95"] Dec 03 19:05:34 crc kubenswrapper[4787]: E1203 19:05:34.687651 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerName="extract-utilities" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.687670 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerName="extract-utilities" Dec 03 19:05:34 crc kubenswrapper[4787]: E1203 19:05:34.687696 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerName="registry-server" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.687705 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerName="registry-server" Dec 03 19:05:34 crc kubenswrapper[4787]: E1203 19:05:34.687734 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerName="extract-content" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.687744 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerName="extract-content" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.688068 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="861cf19a-5812-4179-9eb5-436dfbc561a8" containerName="registry-server" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.690120 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.698823 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nkm95"] Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.702003 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-utilities\") pod \"redhat-operators-nkm95\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.702170 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-catalog-content\") pod \"redhat-operators-nkm95\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.702230 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r8mdt\" (UniqueName: \"kubernetes.io/projected/8b2c4319-fb35-44a1-917a-493bc434f341-kube-api-access-r8mdt\") pod \"redhat-operators-nkm95\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.805275 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-utilities\") pod \"redhat-operators-nkm95\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.805378 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-catalog-content\") pod \"redhat-operators-nkm95\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.805410 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r8mdt\" (UniqueName: \"kubernetes.io/projected/8b2c4319-fb35-44a1-917a-493bc434f341-kube-api-access-r8mdt\") pod \"redhat-operators-nkm95\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.806498 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-catalog-content\") pod \"redhat-operators-nkm95\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.806610 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-utilities\") pod \"redhat-operators-nkm95\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:34 crc kubenswrapper[4787]: I1203 19:05:34.852644 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r8mdt\" (UniqueName: \"kubernetes.io/projected/8b2c4319-fb35-44a1-917a-493bc434f341-kube-api-access-r8mdt\") pod \"redhat-operators-nkm95\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:35 crc kubenswrapper[4787]: I1203 19:05:35.030519 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:35 crc kubenswrapper[4787]: I1203 19:05:35.595262 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-nkm95"] Dec 03 19:05:36 crc kubenswrapper[4787]: I1203 19:05:36.263320 4787 generic.go:334] "Generic (PLEG): container finished" podID="8b2c4319-fb35-44a1-917a-493bc434f341" containerID="adcab57db1e5b7e0998078a3967c49038b47c3416828464fa8c5412cdc9a8aea" exitCode=0 Dec 03 19:05:36 crc kubenswrapper[4787]: I1203 19:05:36.263527 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkm95" event={"ID":"8b2c4319-fb35-44a1-917a-493bc434f341","Type":"ContainerDied","Data":"adcab57db1e5b7e0998078a3967c49038b47c3416828464fa8c5412cdc9a8aea"} Dec 03 19:05:36 crc kubenswrapper[4787]: I1203 19:05:36.263550 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkm95" event={"ID":"8b2c4319-fb35-44a1-917a-493bc434f341","Type":"ContainerStarted","Data":"4d7f7089c9179b853986fdb648e81450c28ca74ad4d2e775158c4c2dfab9bddf"} Dec 03 19:05:37 crc kubenswrapper[4787]: I1203 19:05:37.279171 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkm95" event={"ID":"8b2c4319-fb35-44a1-917a-493bc434f341","Type":"ContainerStarted","Data":"02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d"} Dec 03 19:05:41 crc kubenswrapper[4787]: I1203 19:05:41.325294 4787 generic.go:334] "Generic (PLEG): container finished" podID="8b2c4319-fb35-44a1-917a-493bc434f341" containerID="02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d" exitCode=0 Dec 03 19:05:41 crc kubenswrapper[4787]: I1203 19:05:41.325340 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkm95" event={"ID":"8b2c4319-fb35-44a1-917a-493bc434f341","Type":"ContainerDied","Data":"02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d"} Dec 03 19:05:42 crc kubenswrapper[4787]: I1203 19:05:42.339995 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkm95" event={"ID":"8b2c4319-fb35-44a1-917a-493bc434f341","Type":"ContainerStarted","Data":"da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859"} Dec 03 19:05:42 crc kubenswrapper[4787]: I1203 19:05:42.365086 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-nkm95" podStartSLOduration=2.875750373 podStartE2EDuration="8.365060752s" podCreationTimestamp="2025-12-03 19:05:34 +0000 UTC" firstStartedPulling="2025-12-03 19:05:36.267217383 +0000 UTC m=+6753.084688332" lastFinishedPulling="2025-12-03 19:05:41.756527752 +0000 UTC m=+6758.573998711" observedRunningTime="2025-12-03 19:05:42.363038219 +0000 UTC m=+6759.180509188" watchObservedRunningTime="2025-12-03 19:05:42.365060752 +0000 UTC m=+6759.182531751" Dec 03 19:05:45 crc kubenswrapper[4787]: I1203 19:05:45.031217 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:45 crc kubenswrapper[4787]: I1203 19:05:45.032346 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:05:45 crc kubenswrapper[4787]: I1203 19:05:45.767244 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:05:45 crc kubenswrapper[4787]: E1203 19:05:45.768413 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:05:46 crc kubenswrapper[4787]: I1203 19:05:46.107564 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nkm95" podUID="8b2c4319-fb35-44a1-917a-493bc434f341" containerName="registry-server" probeResult="failure" output=< Dec 03 19:05:46 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 19:05:46 crc kubenswrapper[4787]: > Dec 03 19:05:56 crc kubenswrapper[4787]: I1203 19:05:56.118744 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-nkm95" podUID="8b2c4319-fb35-44a1-917a-493bc434f341" containerName="registry-server" probeResult="failure" output=< Dec 03 19:05:56 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 19:05:56 crc kubenswrapper[4787]: > Dec 03 19:05:57 crc kubenswrapper[4787]: I1203 19:05:57.773762 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:05:57 crc kubenswrapper[4787]: E1203 19:05:57.774884 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:06:05 crc kubenswrapper[4787]: I1203 19:06:05.105106 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:06:05 crc kubenswrapper[4787]: I1203 19:06:05.181206 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:06:05 crc kubenswrapper[4787]: I1203 19:06:05.904966 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nkm95"] Dec 03 19:06:06 crc kubenswrapper[4787]: I1203 19:06:06.695620 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-nkm95" podUID="8b2c4319-fb35-44a1-917a-493bc434f341" containerName="registry-server" containerID="cri-o://da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859" gracePeriod=2 Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.372214 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.493832 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-catalog-content\") pod \"8b2c4319-fb35-44a1-917a-493bc434f341\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.493968 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-utilities\") pod \"8b2c4319-fb35-44a1-917a-493bc434f341\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.494130 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r8mdt\" (UniqueName: \"kubernetes.io/projected/8b2c4319-fb35-44a1-917a-493bc434f341-kube-api-access-r8mdt\") pod \"8b2c4319-fb35-44a1-917a-493bc434f341\" (UID: \"8b2c4319-fb35-44a1-917a-493bc434f341\") " Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.495009 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-utilities" (OuterVolumeSpecName: "utilities") pod "8b2c4319-fb35-44a1-917a-493bc434f341" (UID: "8b2c4319-fb35-44a1-917a-493bc434f341"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.502521 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8b2c4319-fb35-44a1-917a-493bc434f341-kube-api-access-r8mdt" (OuterVolumeSpecName: "kube-api-access-r8mdt") pod "8b2c4319-fb35-44a1-917a-493bc434f341" (UID: "8b2c4319-fb35-44a1-917a-493bc434f341"). InnerVolumeSpecName "kube-api-access-r8mdt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.588816 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8b2c4319-fb35-44a1-917a-493bc434f341" (UID: "8b2c4319-fb35-44a1-917a-493bc434f341"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.597854 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r8mdt\" (UniqueName: \"kubernetes.io/projected/8b2c4319-fb35-44a1-917a-493bc434f341-kube-api-access-r8mdt\") on node \"crc\" DevicePath \"\"" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.597879 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.597890 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8b2c4319-fb35-44a1-917a-493bc434f341-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.709604 4787 generic.go:334] "Generic (PLEG): container finished" podID="8b2c4319-fb35-44a1-917a-493bc434f341" containerID="da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859" exitCode=0 Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.709676 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkm95" event={"ID":"8b2c4319-fb35-44a1-917a-493bc434f341","Type":"ContainerDied","Data":"da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859"} Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.709744 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-nkm95" event={"ID":"8b2c4319-fb35-44a1-917a-493bc434f341","Type":"ContainerDied","Data":"4d7f7089c9179b853986fdb648e81450c28ca74ad4d2e775158c4c2dfab9bddf"} Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.709770 4787 scope.go:117] "RemoveContainer" containerID="da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.709829 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-nkm95" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.733777 4787 scope.go:117] "RemoveContainer" containerID="02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.764264 4787 scope.go:117] "RemoveContainer" containerID="adcab57db1e5b7e0998078a3967c49038b47c3416828464fa8c5412cdc9a8aea" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.781309 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-nkm95"] Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.799290 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-nkm95"] Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.827785 4787 scope.go:117] "RemoveContainer" containerID="da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859" Dec 03 19:06:07 crc kubenswrapper[4787]: E1203 19:06:07.828405 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859\": container with ID starting with da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859 not found: ID does not exist" containerID="da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.828467 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859"} err="failed to get container status \"da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859\": rpc error: code = NotFound desc = could not find container \"da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859\": container with ID starting with da02eb6b5a0c738a06da5f20793b1616360c85d9044868acdc13d8ad4e8a2859 not found: ID does not exist" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.828500 4787 scope.go:117] "RemoveContainer" containerID="02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d" Dec 03 19:06:07 crc kubenswrapper[4787]: E1203 19:06:07.829123 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d\": container with ID starting with 02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d not found: ID does not exist" containerID="02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.829152 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d"} err="failed to get container status \"02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d\": rpc error: code = NotFound desc = could not find container \"02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d\": container with ID starting with 02c63dc0818b6877303ff1002f4889506c99bcbf7973805a0898596f5637cf0d not found: ID does not exist" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.829173 4787 scope.go:117] "RemoveContainer" containerID="adcab57db1e5b7e0998078a3967c49038b47c3416828464fa8c5412cdc9a8aea" Dec 03 19:06:07 crc kubenswrapper[4787]: E1203 19:06:07.829771 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"adcab57db1e5b7e0998078a3967c49038b47c3416828464fa8c5412cdc9a8aea\": container with ID starting with adcab57db1e5b7e0998078a3967c49038b47c3416828464fa8c5412cdc9a8aea not found: ID does not exist" containerID="adcab57db1e5b7e0998078a3967c49038b47c3416828464fa8c5412cdc9a8aea" Dec 03 19:06:07 crc kubenswrapper[4787]: I1203 19:06:07.829833 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"adcab57db1e5b7e0998078a3967c49038b47c3416828464fa8c5412cdc9a8aea"} err="failed to get container status \"adcab57db1e5b7e0998078a3967c49038b47c3416828464fa8c5412cdc9a8aea\": rpc error: code = NotFound desc = could not find container \"adcab57db1e5b7e0998078a3967c49038b47c3416828464fa8c5412cdc9a8aea\": container with ID starting with adcab57db1e5b7e0998078a3967c49038b47c3416828464fa8c5412cdc9a8aea not found: ID does not exist" Dec 03 19:06:08 crc kubenswrapper[4787]: I1203 19:06:08.766170 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:06:08 crc kubenswrapper[4787]: E1203 19:06:08.766681 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:06:09 crc kubenswrapper[4787]: I1203 19:06:09.789560 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8b2c4319-fb35-44a1-917a-493bc434f341" path="/var/lib/kubelet/pods/8b2c4319-fb35-44a1-917a-493bc434f341/volumes" Dec 03 19:06:20 crc kubenswrapper[4787]: I1203 19:06:20.795197 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:06:20 crc kubenswrapper[4787]: E1203 19:06:20.796701 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:06:31 crc kubenswrapper[4787]: I1203 19:06:31.766550 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:06:31 crc kubenswrapper[4787]: E1203 19:06:31.767757 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:06:44 crc kubenswrapper[4787]: I1203 19:06:44.766727 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:06:44 crc kubenswrapper[4787]: E1203 19:06:44.768780 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:06:59 crc kubenswrapper[4787]: I1203 19:06:59.766619 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:06:59 crc kubenswrapper[4787]: E1203 19:06:59.767363 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:07:14 crc kubenswrapper[4787]: I1203 19:07:14.766310 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:07:14 crc kubenswrapper[4787]: E1203 19:07:14.767306 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:07:29 crc kubenswrapper[4787]: I1203 19:07:29.767852 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:07:29 crc kubenswrapper[4787]: E1203 19:07:29.769199 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:07:41 crc kubenswrapper[4787]: I1203 19:07:41.766639 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:07:41 crc kubenswrapper[4787]: E1203 19:07:41.768645 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:07:53 crc kubenswrapper[4787]: I1203 19:07:53.780069 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:07:53 crc kubenswrapper[4787]: E1203 19:07:53.780935 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:08:05 crc kubenswrapper[4787]: I1203 19:08:05.767196 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:08:05 crc kubenswrapper[4787]: E1203 19:08:05.768518 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:08:20 crc kubenswrapper[4787]: I1203 19:08:20.766538 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:08:20 crc kubenswrapper[4787]: E1203 19:08:20.767634 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:08:34 crc kubenswrapper[4787]: I1203 19:08:34.766945 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:08:34 crc kubenswrapper[4787]: E1203 19:08:34.767878 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:08:45 crc kubenswrapper[4787]: I1203 19:08:45.765947 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:08:45 crc kubenswrapper[4787]: E1203 19:08:45.766634 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:08:58 crc kubenswrapper[4787]: I1203 19:08:58.766390 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:08:58 crc kubenswrapper[4787]: E1203 19:08:58.767510 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:09:11 crc kubenswrapper[4787]: I1203 19:09:11.766787 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:09:11 crc kubenswrapper[4787]: E1203 19:09:11.769379 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:09:23 crc kubenswrapper[4787]: I1203 19:09:23.783501 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:09:23 crc kubenswrapper[4787]: E1203 19:09:23.784891 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:09:36 crc kubenswrapper[4787]: I1203 19:09:36.766690 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:09:36 crc kubenswrapper[4787]: E1203 19:09:36.767608 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:09:48 crc kubenswrapper[4787]: I1203 19:09:48.766760 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:09:48 crc kubenswrapper[4787]: E1203 19:09:48.767811 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.414229 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-n256q"] Dec 03 19:09:56 crc kubenswrapper[4787]: E1203 19:09:56.415297 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b2c4319-fb35-44a1-917a-493bc434f341" containerName="extract-content" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.415314 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b2c4319-fb35-44a1-917a-493bc434f341" containerName="extract-content" Dec 03 19:09:56 crc kubenswrapper[4787]: E1203 19:09:56.415355 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b2c4319-fb35-44a1-917a-493bc434f341" containerName="registry-server" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.415363 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b2c4319-fb35-44a1-917a-493bc434f341" containerName="registry-server" Dec 03 19:09:56 crc kubenswrapper[4787]: E1203 19:09:56.415397 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8b2c4319-fb35-44a1-917a-493bc434f341" containerName="extract-utilities" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.415406 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="8b2c4319-fb35-44a1-917a-493bc434f341" containerName="extract-utilities" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.415692 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="8b2c4319-fb35-44a1-917a-493bc434f341" containerName="registry-server" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.422174 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.500910 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n256q"] Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.591773 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-catalog-content\") pod \"redhat-marketplace-n256q\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.591850 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-utilities\") pod \"redhat-marketplace-n256q\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.591996 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt5vc\" (UniqueName: \"kubernetes.io/projected/f835b056-8d2a-4f8f-820c-ded31737c5c1-kube-api-access-nt5vc\") pod \"redhat-marketplace-n256q\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.694079 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-utilities\") pod \"redhat-marketplace-n256q\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.694252 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt5vc\" (UniqueName: \"kubernetes.io/projected/f835b056-8d2a-4f8f-820c-ded31737c5c1-kube-api-access-nt5vc\") pod \"redhat-marketplace-n256q\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.694406 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-catalog-content\") pod \"redhat-marketplace-n256q\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.694526 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-utilities\") pod \"redhat-marketplace-n256q\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.694892 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-catalog-content\") pod \"redhat-marketplace-n256q\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.725227 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt5vc\" (UniqueName: \"kubernetes.io/projected/f835b056-8d2a-4f8f-820c-ded31737c5c1-kube-api-access-nt5vc\") pod \"redhat-marketplace-n256q\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:09:56 crc kubenswrapper[4787]: I1203 19:09:56.777953 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:09:57 crc kubenswrapper[4787]: I1203 19:09:57.308510 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-n256q"] Dec 03 19:09:57 crc kubenswrapper[4787]: I1203 19:09:57.867119 4787 generic.go:334] "Generic (PLEG): container finished" podID="f835b056-8d2a-4f8f-820c-ded31737c5c1" containerID="581102acf9c1276bf8ac81f730ff8fb664a7b1324341e191ce7d1c10bda37eda" exitCode=0 Dec 03 19:09:57 crc kubenswrapper[4787]: I1203 19:09:57.867287 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n256q" event={"ID":"f835b056-8d2a-4f8f-820c-ded31737c5c1","Type":"ContainerDied","Data":"581102acf9c1276bf8ac81f730ff8fb664a7b1324341e191ce7d1c10bda37eda"} Dec 03 19:09:57 crc kubenswrapper[4787]: I1203 19:09:57.867462 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n256q" event={"ID":"f835b056-8d2a-4f8f-820c-ded31737c5c1","Type":"ContainerStarted","Data":"86be874d208acbd094790e5a540611ff8096aa7eff4b75eb2f1a550a77dc174d"} Dec 03 19:09:57 crc kubenswrapper[4787]: I1203 19:09:57.870283 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:09:58 crc kubenswrapper[4787]: I1203 19:09:58.879101 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n256q" event={"ID":"f835b056-8d2a-4f8f-820c-ded31737c5c1","Type":"ContainerStarted","Data":"bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233"} Dec 03 19:09:59 crc kubenswrapper[4787]: I1203 19:09:59.895424 4787 generic.go:334] "Generic (PLEG): container finished" podID="f835b056-8d2a-4f8f-820c-ded31737c5c1" containerID="bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233" exitCode=0 Dec 03 19:09:59 crc kubenswrapper[4787]: I1203 19:09:59.895569 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n256q" event={"ID":"f835b056-8d2a-4f8f-820c-ded31737c5c1","Type":"ContainerDied","Data":"bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233"} Dec 03 19:10:00 crc kubenswrapper[4787]: I1203 19:10:00.908930 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n256q" event={"ID":"f835b056-8d2a-4f8f-820c-ded31737c5c1","Type":"ContainerStarted","Data":"1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f"} Dec 03 19:10:00 crc kubenswrapper[4787]: I1203 19:10:00.934982 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-n256q" podStartSLOduration=2.49258761 podStartE2EDuration="4.934964248s" podCreationTimestamp="2025-12-03 19:09:56 +0000 UTC" firstStartedPulling="2025-12-03 19:09:57.869505165 +0000 UTC m=+7014.686976134" lastFinishedPulling="2025-12-03 19:10:00.311881813 +0000 UTC m=+7017.129352772" observedRunningTime="2025-12-03 19:10:00.932294117 +0000 UTC m=+7017.749765086" watchObservedRunningTime="2025-12-03 19:10:00.934964248 +0000 UTC m=+7017.752435207" Dec 03 19:10:01 crc kubenswrapper[4787]: I1203 19:10:01.766080 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:10:01 crc kubenswrapper[4787]: E1203 19:10:01.766396 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:10:06 crc kubenswrapper[4787]: I1203 19:10:06.778552 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:10:06 crc kubenswrapper[4787]: I1203 19:10:06.779242 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:10:06 crc kubenswrapper[4787]: I1203 19:10:06.853513 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:10:07 crc kubenswrapper[4787]: I1203 19:10:07.052223 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:10:07 crc kubenswrapper[4787]: I1203 19:10:07.105715 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n256q"] Dec 03 19:10:08 crc kubenswrapper[4787]: I1203 19:10:08.997358 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-n256q" podUID="f835b056-8d2a-4f8f-820c-ded31737c5c1" containerName="registry-server" containerID="cri-o://1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f" gracePeriod=2 Dec 03 19:10:09 crc kubenswrapper[4787]: I1203 19:10:09.576146 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:10:09 crc kubenswrapper[4787]: I1203 19:10:09.669752 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-catalog-content\") pod \"f835b056-8d2a-4f8f-820c-ded31737c5c1\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " Dec 03 19:10:09 crc kubenswrapper[4787]: I1203 19:10:09.669846 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-utilities\") pod \"f835b056-8d2a-4f8f-820c-ded31737c5c1\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " Dec 03 19:10:09 crc kubenswrapper[4787]: I1203 19:10:09.670046 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nt5vc\" (UniqueName: \"kubernetes.io/projected/f835b056-8d2a-4f8f-820c-ded31737c5c1-kube-api-access-nt5vc\") pod \"f835b056-8d2a-4f8f-820c-ded31737c5c1\" (UID: \"f835b056-8d2a-4f8f-820c-ded31737c5c1\") " Dec 03 19:10:09 crc kubenswrapper[4787]: I1203 19:10:09.670680 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-utilities" (OuterVolumeSpecName: "utilities") pod "f835b056-8d2a-4f8f-820c-ded31737c5c1" (UID: "f835b056-8d2a-4f8f-820c-ded31737c5c1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:10:09 crc kubenswrapper[4787]: I1203 19:10:09.678379 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f835b056-8d2a-4f8f-820c-ded31737c5c1-kube-api-access-nt5vc" (OuterVolumeSpecName: "kube-api-access-nt5vc") pod "f835b056-8d2a-4f8f-820c-ded31737c5c1" (UID: "f835b056-8d2a-4f8f-820c-ded31737c5c1"). InnerVolumeSpecName "kube-api-access-nt5vc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:10:09 crc kubenswrapper[4787]: I1203 19:10:09.686600 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f835b056-8d2a-4f8f-820c-ded31737c5c1" (UID: "f835b056-8d2a-4f8f-820c-ded31737c5c1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:10:09 crc kubenswrapper[4787]: I1203 19:10:09.771939 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nt5vc\" (UniqueName: \"kubernetes.io/projected/f835b056-8d2a-4f8f-820c-ded31737c5c1-kube-api-access-nt5vc\") on node \"crc\" DevicePath \"\"" Dec 03 19:10:09 crc kubenswrapper[4787]: I1203 19:10:09.772215 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:10:09 crc kubenswrapper[4787]: I1203 19:10:09.772224 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f835b056-8d2a-4f8f-820c-ded31737c5c1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.009089 4787 generic.go:334] "Generic (PLEG): container finished" podID="f835b056-8d2a-4f8f-820c-ded31737c5c1" containerID="1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f" exitCode=0 Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.009127 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n256q" event={"ID":"f835b056-8d2a-4f8f-820c-ded31737c5c1","Type":"ContainerDied","Data":"1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f"} Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.009161 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-n256q" event={"ID":"f835b056-8d2a-4f8f-820c-ded31737c5c1","Type":"ContainerDied","Data":"86be874d208acbd094790e5a540611ff8096aa7eff4b75eb2f1a550a77dc174d"} Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.009178 4787 scope.go:117] "RemoveContainer" containerID="1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f" Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.009195 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-n256q" Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.054820 4787 scope.go:117] "RemoveContainer" containerID="bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233" Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.063188 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-n256q"] Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.077561 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-n256q"] Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.115223 4787 scope.go:117] "RemoveContainer" containerID="581102acf9c1276bf8ac81f730ff8fb664a7b1324341e191ce7d1c10bda37eda" Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.145267 4787 scope.go:117] "RemoveContainer" containerID="1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f" Dec 03 19:10:10 crc kubenswrapper[4787]: E1203 19:10:10.145687 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f\": container with ID starting with 1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f not found: ID does not exist" containerID="1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f" Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.145724 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f"} err="failed to get container status \"1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f\": rpc error: code = NotFound desc = could not find container \"1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f\": container with ID starting with 1f131b548d1e46c1dee7d426354a0cebc2d811014f4b1caceef449b2549f732f not found: ID does not exist" Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.145748 4787 scope.go:117] "RemoveContainer" containerID="bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233" Dec 03 19:10:10 crc kubenswrapper[4787]: E1203 19:10:10.145987 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233\": container with ID starting with bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233 not found: ID does not exist" containerID="bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233" Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.146044 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233"} err="failed to get container status \"bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233\": rpc error: code = NotFound desc = could not find container \"bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233\": container with ID starting with bee82fa354d201f86308ec8473959fc9d0d721e9506fbaf8a51f497385bed233 not found: ID does not exist" Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.146065 4787 scope.go:117] "RemoveContainer" containerID="581102acf9c1276bf8ac81f730ff8fb664a7b1324341e191ce7d1c10bda37eda" Dec 03 19:10:10 crc kubenswrapper[4787]: E1203 19:10:10.146309 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"581102acf9c1276bf8ac81f730ff8fb664a7b1324341e191ce7d1c10bda37eda\": container with ID starting with 581102acf9c1276bf8ac81f730ff8fb664a7b1324341e191ce7d1c10bda37eda not found: ID does not exist" containerID="581102acf9c1276bf8ac81f730ff8fb664a7b1324341e191ce7d1c10bda37eda" Dec 03 19:10:10 crc kubenswrapper[4787]: I1203 19:10:10.146339 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"581102acf9c1276bf8ac81f730ff8fb664a7b1324341e191ce7d1c10bda37eda"} err="failed to get container status \"581102acf9c1276bf8ac81f730ff8fb664a7b1324341e191ce7d1c10bda37eda\": rpc error: code = NotFound desc = could not find container \"581102acf9c1276bf8ac81f730ff8fb664a7b1324341e191ce7d1c10bda37eda\": container with ID starting with 581102acf9c1276bf8ac81f730ff8fb664a7b1324341e191ce7d1c10bda37eda not found: ID does not exist" Dec 03 19:10:11 crc kubenswrapper[4787]: I1203 19:10:11.788325 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f835b056-8d2a-4f8f-820c-ded31737c5c1" path="/var/lib/kubelet/pods/f835b056-8d2a-4f8f-820c-ded31737c5c1/volumes" Dec 03 19:10:14 crc kubenswrapper[4787]: I1203 19:10:14.767074 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:10:14 crc kubenswrapper[4787]: E1203 19:10:14.768218 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:10:25 crc kubenswrapper[4787]: I1203 19:10:25.766386 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:10:26 crc kubenswrapper[4787]: I1203 19:10:26.245811 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"65214d10d37779bf02fb7aef42e332ae9add4645b6632d81a6d49b8d359c7917"} Dec 03 19:12:48 crc kubenswrapper[4787]: I1203 19:12:48.990136 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:12:48 crc kubenswrapper[4787]: I1203 19:12:48.990720 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:13:18 crc kubenswrapper[4787]: I1203 19:13:18.990540 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:13:18 crc kubenswrapper[4787]: I1203 19:13:18.991357 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:13:48 crc kubenswrapper[4787]: I1203 19:13:48.990734 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:13:48 crc kubenswrapper[4787]: I1203 19:13:48.991739 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:13:48 crc kubenswrapper[4787]: I1203 19:13:48.991827 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 19:13:48 crc kubenswrapper[4787]: I1203 19:13:48.993145 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"65214d10d37779bf02fb7aef42e332ae9add4645b6632d81a6d49b8d359c7917"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:13:48 crc kubenswrapper[4787]: I1203 19:13:48.993252 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://65214d10d37779bf02fb7aef42e332ae9add4645b6632d81a6d49b8d359c7917" gracePeriod=600 Dec 03 19:13:50 crc kubenswrapper[4787]: I1203 19:13:50.106957 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="65214d10d37779bf02fb7aef42e332ae9add4645b6632d81a6d49b8d359c7917" exitCode=0 Dec 03 19:13:50 crc kubenswrapper[4787]: I1203 19:13:50.107863 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"65214d10d37779bf02fb7aef42e332ae9add4645b6632d81a6d49b8d359c7917"} Dec 03 19:13:50 crc kubenswrapper[4787]: I1203 19:13:50.107905 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514"} Dec 03 19:13:50 crc kubenswrapper[4787]: I1203 19:13:50.107930 4787 scope.go:117] "RemoveContainer" containerID="1e4f4f44a821826f4aa3e55c137086afb4c837a2fa2fd7b19d25b556aec14e28" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.192140 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4"] Dec 03 19:15:00 crc kubenswrapper[4787]: E1203 19:15:00.193273 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f835b056-8d2a-4f8f-820c-ded31737c5c1" containerName="registry-server" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.193293 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f835b056-8d2a-4f8f-820c-ded31737c5c1" containerName="registry-server" Dec 03 19:15:00 crc kubenswrapper[4787]: E1203 19:15:00.193332 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f835b056-8d2a-4f8f-820c-ded31737c5c1" containerName="extract-utilities" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.193341 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f835b056-8d2a-4f8f-820c-ded31737c5c1" containerName="extract-utilities" Dec 03 19:15:00 crc kubenswrapper[4787]: E1203 19:15:00.193355 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f835b056-8d2a-4f8f-820c-ded31737c5c1" containerName="extract-content" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.193364 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f835b056-8d2a-4f8f-820c-ded31737c5c1" containerName="extract-content" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.193634 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f835b056-8d2a-4f8f-820c-ded31737c5c1" containerName="registry-server" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.194566 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.205740 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.205746 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.227321 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4"] Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.287501 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7m9h\" (UniqueName: \"kubernetes.io/projected/e38b6288-b87a-4bc4-913b-368f2c86739a-kube-api-access-m7m9h\") pod \"collect-profiles-29413155-l8cq4\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.287619 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e38b6288-b87a-4bc4-913b-368f2c86739a-config-volume\") pod \"collect-profiles-29413155-l8cq4\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.287720 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e38b6288-b87a-4bc4-913b-368f2c86739a-secret-volume\") pod \"collect-profiles-29413155-l8cq4\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.389885 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7m9h\" (UniqueName: \"kubernetes.io/projected/e38b6288-b87a-4bc4-913b-368f2c86739a-kube-api-access-m7m9h\") pod \"collect-profiles-29413155-l8cq4\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.389979 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e38b6288-b87a-4bc4-913b-368f2c86739a-config-volume\") pod \"collect-profiles-29413155-l8cq4\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.390085 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e38b6288-b87a-4bc4-913b-368f2c86739a-secret-volume\") pod \"collect-profiles-29413155-l8cq4\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.390751 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e38b6288-b87a-4bc4-913b-368f2c86739a-config-volume\") pod \"collect-profiles-29413155-l8cq4\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.396445 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e38b6288-b87a-4bc4-913b-368f2c86739a-secret-volume\") pod \"collect-profiles-29413155-l8cq4\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.426352 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7m9h\" (UniqueName: \"kubernetes.io/projected/e38b6288-b87a-4bc4-913b-368f2c86739a-kube-api-access-m7m9h\") pod \"collect-profiles-29413155-l8cq4\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:00 crc kubenswrapper[4787]: I1203 19:15:00.533246 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:01 crc kubenswrapper[4787]: I1203 19:15:01.051278 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4"] Dec 03 19:15:02 crc kubenswrapper[4787]: I1203 19:15:02.052965 4787 generic.go:334] "Generic (PLEG): container finished" podID="e38b6288-b87a-4bc4-913b-368f2c86739a" containerID="0554e832dee691f9ace161ff3c35bcb042bf756eace3732013305dfb72767f80" exitCode=0 Dec 03 19:15:02 crc kubenswrapper[4787]: I1203 19:15:02.053043 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" event={"ID":"e38b6288-b87a-4bc4-913b-368f2c86739a","Type":"ContainerDied","Data":"0554e832dee691f9ace161ff3c35bcb042bf756eace3732013305dfb72767f80"} Dec 03 19:15:02 crc kubenswrapper[4787]: I1203 19:15:02.053252 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" event={"ID":"e38b6288-b87a-4bc4-913b-368f2c86739a","Type":"ContainerStarted","Data":"17f8cd8c2cc7e5db5c49cbf7663cc452ac345066258872b63223da6f82163ffe"} Dec 03 19:15:03 crc kubenswrapper[4787]: I1203 19:15:03.519661 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:03 crc kubenswrapper[4787]: I1203 19:15:03.576217 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7m9h\" (UniqueName: \"kubernetes.io/projected/e38b6288-b87a-4bc4-913b-368f2c86739a-kube-api-access-m7m9h\") pod \"e38b6288-b87a-4bc4-913b-368f2c86739a\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " Dec 03 19:15:03 crc kubenswrapper[4787]: I1203 19:15:03.576305 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e38b6288-b87a-4bc4-913b-368f2c86739a-secret-volume\") pod \"e38b6288-b87a-4bc4-913b-368f2c86739a\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " Dec 03 19:15:03 crc kubenswrapper[4787]: I1203 19:15:03.576410 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e38b6288-b87a-4bc4-913b-368f2c86739a-config-volume\") pod \"e38b6288-b87a-4bc4-913b-368f2c86739a\" (UID: \"e38b6288-b87a-4bc4-913b-368f2c86739a\") " Dec 03 19:15:03 crc kubenswrapper[4787]: I1203 19:15:03.578114 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e38b6288-b87a-4bc4-913b-368f2c86739a-config-volume" (OuterVolumeSpecName: "config-volume") pod "e38b6288-b87a-4bc4-913b-368f2c86739a" (UID: "e38b6288-b87a-4bc4-913b-368f2c86739a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:15:03 crc kubenswrapper[4787]: I1203 19:15:03.585359 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e38b6288-b87a-4bc4-913b-368f2c86739a-kube-api-access-m7m9h" (OuterVolumeSpecName: "kube-api-access-m7m9h") pod "e38b6288-b87a-4bc4-913b-368f2c86739a" (UID: "e38b6288-b87a-4bc4-913b-368f2c86739a"). InnerVolumeSpecName "kube-api-access-m7m9h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:15:03 crc kubenswrapper[4787]: I1203 19:15:03.585430 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e38b6288-b87a-4bc4-913b-368f2c86739a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e38b6288-b87a-4bc4-913b-368f2c86739a" (UID: "e38b6288-b87a-4bc4-913b-368f2c86739a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:15:03 crc kubenswrapper[4787]: I1203 19:15:03.679028 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7m9h\" (UniqueName: \"kubernetes.io/projected/e38b6288-b87a-4bc4-913b-368f2c86739a-kube-api-access-m7m9h\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:03 crc kubenswrapper[4787]: I1203 19:15:03.679067 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e38b6288-b87a-4bc4-913b-368f2c86739a-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:03 crc kubenswrapper[4787]: I1203 19:15:03.679080 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e38b6288-b87a-4bc4-913b-368f2c86739a-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:04 crc kubenswrapper[4787]: I1203 19:15:04.073684 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" event={"ID":"e38b6288-b87a-4bc4-913b-368f2c86739a","Type":"ContainerDied","Data":"17f8cd8c2cc7e5db5c49cbf7663cc452ac345066258872b63223da6f82163ffe"} Dec 03 19:15:04 crc kubenswrapper[4787]: I1203 19:15:04.074122 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="17f8cd8c2cc7e5db5c49cbf7663cc452ac345066258872b63223da6f82163ffe" Dec 03 19:15:04 crc kubenswrapper[4787]: I1203 19:15:04.073749 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413155-l8cq4" Dec 03 19:15:04 crc kubenswrapper[4787]: I1203 19:15:04.628726 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs"] Dec 03 19:15:04 crc kubenswrapper[4787]: I1203 19:15:04.645152 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413110-pdmfs"] Dec 03 19:15:05 crc kubenswrapper[4787]: I1203 19:15:05.778788 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="356d1186-2f36-4101-80d0-0d240814bb08" path="/var/lib/kubelet/pods/356d1186-2f36-4101-80d0-0d240814bb08/volumes" Dec 03 19:15:09 crc kubenswrapper[4787]: E1203 19:15:09.857796 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice/crio-17f8cd8c2cc7e5db5c49cbf7663cc452ac345066258872b63223da6f82163ffe\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice\": RecentStats: unable to find data in memory cache]" Dec 03 19:15:20 crc kubenswrapper[4787]: E1203 19:15:20.146102 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice/crio-17f8cd8c2cc7e5db5c49cbf7663cc452ac345066258872b63223da6f82163ffe\": RecentStats: unable to find data in memory cache]" Dec 03 19:15:20 crc kubenswrapper[4787]: I1203 19:15:20.626139 4787 generic.go:334] "Generic (PLEG): container finished" podID="4c367f96-011e-4a7c-89e6-53c9ed2d1c90" containerID="883f6485c880565394bf58ea6259a9a888ecffce140118ca938c07df4197e362" exitCode=0 Dec 03 19:15:20 crc kubenswrapper[4787]: I1203 19:15:20.626225 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4c367f96-011e-4a7c-89e6-53c9ed2d1c90","Type":"ContainerDied","Data":"883f6485c880565394bf58ea6259a9a888ecffce140118ca938c07df4197e362"} Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.082737 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.151488 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nqwln\" (UniqueName: \"kubernetes.io/projected/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-kube-api-access-nqwln\") pod \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.151850 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config-secret\") pod \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.152131 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ca-certs\") pod \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.152812 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-temporary\") pod \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.153045 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.153169 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ssh-key\") pod \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.153259 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config\") pod \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.153334 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-workdir\") pod \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.153416 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-config-data\") pod \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\" (UID: \"4c367f96-011e-4a7c-89e6-53c9ed2d1c90\") " Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.156086 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "4c367f96-011e-4a7c-89e6-53c9ed2d1c90" (UID: "4c367f96-011e-4a7c-89e6-53c9ed2d1c90"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.161631 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-kube-api-access-nqwln" (OuterVolumeSpecName: "kube-api-access-nqwln") pod "4c367f96-011e-4a7c-89e6-53c9ed2d1c90" (UID: "4c367f96-011e-4a7c-89e6-53c9ed2d1c90"). InnerVolumeSpecName "kube-api-access-nqwln". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.162119 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "test-operator-logs") pod "4c367f96-011e-4a7c-89e6-53c9ed2d1c90" (UID: "4c367f96-011e-4a7c-89e6-53c9ed2d1c90"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.164295 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "4c367f96-011e-4a7c-89e6-53c9ed2d1c90" (UID: "4c367f96-011e-4a7c-89e6-53c9ed2d1c90"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.166978 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-config-data" (OuterVolumeSpecName: "config-data") pod "4c367f96-011e-4a7c-89e6-53c9ed2d1c90" (UID: "4c367f96-011e-4a7c-89e6-53c9ed2d1c90"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.207958 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "4c367f96-011e-4a7c-89e6-53c9ed2d1c90" (UID: "4c367f96-011e-4a7c-89e6-53c9ed2d1c90"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.209335 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "4c367f96-011e-4a7c-89e6-53c9ed2d1c90" (UID: "4c367f96-011e-4a7c-89e6-53c9ed2d1c90"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.220898 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "4c367f96-011e-4a7c-89e6-53c9ed2d1c90" (UID: "4c367f96-011e-4a7c-89e6-53c9ed2d1c90"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.222061 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4c367f96-011e-4a7c-89e6-53c9ed2d1c90" (UID: "4c367f96-011e-4a7c-89e6-53c9ed2d1c90"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.258382 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nqwln\" (UniqueName: \"kubernetes.io/projected/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-kube-api-access-nqwln\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.258432 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.258443 4787 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.258453 4787 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.258491 4787 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.258501 4787 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.258512 4787 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.258521 4787 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.258530 4787 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4c367f96-011e-4a7c-89e6-53c9ed2d1c90-config-data\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.314934 4787 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.361093 4787 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.649569 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest" event={"ID":"4c367f96-011e-4a7c-89e6-53c9ed2d1c90","Type":"ContainerDied","Data":"50730d92b8a739cb20322926f7d6b51e7e525cbf60f5bc6bfa1fbef53c263f0c"} Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.649606 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="50730d92b8a739cb20322926f7d6b51e7e525cbf60f5bc6bfa1fbef53c263f0c" Dec 03 19:15:22 crc kubenswrapper[4787]: I1203 19:15:22.649663 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.504310 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 19:15:29 crc kubenswrapper[4787]: E1203 19:15:29.505260 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c367f96-011e-4a7c-89e6-53c9ed2d1c90" containerName="tempest-tests-tempest-tests-runner" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.505273 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c367f96-011e-4a7c-89e6-53c9ed2d1c90" containerName="tempest-tests-tempest-tests-runner" Dec 03 19:15:29 crc kubenswrapper[4787]: E1203 19:15:29.505305 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e38b6288-b87a-4bc4-913b-368f2c86739a" containerName="collect-profiles" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.505311 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e38b6288-b87a-4bc4-913b-368f2c86739a" containerName="collect-profiles" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.505510 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e38b6288-b87a-4bc4-913b-368f2c86739a" containerName="collect-profiles" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.505540 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c367f96-011e-4a7c-89e6-53c9ed2d1c90" containerName="tempest-tests-tempest-tests-runner" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.506434 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.508802 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-mn9nq" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.520226 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.654423 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qpltq\" (UniqueName: \"kubernetes.io/projected/dba0eca5-4008-4ce5-b555-d6db725f6466-kube-api-access-qpltq\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dba0eca5-4008-4ce5-b555-d6db725f6466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.654558 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dba0eca5-4008-4ce5-b555-d6db725f6466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.756981 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qpltq\" (UniqueName: \"kubernetes.io/projected/dba0eca5-4008-4ce5-b555-d6db725f6466-kube-api-access-qpltq\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dba0eca5-4008-4ce5-b555-d6db725f6466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.757430 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dba0eca5-4008-4ce5-b555-d6db725f6466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.759014 4787 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dba0eca5-4008-4ce5-b555-d6db725f6466\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.807843 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dba0eca5-4008-4ce5-b555-d6db725f6466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.815816 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qpltq\" (UniqueName: \"kubernetes.io/projected/dba0eca5-4008-4ce5-b555-d6db725f6466-kube-api-access-qpltq\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"dba0eca5-4008-4ce5-b555-d6db725f6466\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 19:15:29 crc kubenswrapper[4787]: I1203 19:15:29.830470 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 03 19:15:30 crc kubenswrapper[4787]: I1203 19:15:30.329223 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 03 19:15:30 crc kubenswrapper[4787]: W1203 19:15:30.338377 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddba0eca5_4008_4ce5_b555_d6db725f6466.slice/crio-6827790696bf4e10023f202eb462ce7115c3fc12397d20fb37d89ffe03f18ea6 WatchSource:0}: Error finding container 6827790696bf4e10023f202eb462ce7115c3fc12397d20fb37d89ffe03f18ea6: Status 404 returned error can't find the container with id 6827790696bf4e10023f202eb462ce7115c3fc12397d20fb37d89ffe03f18ea6 Dec 03 19:15:30 crc kubenswrapper[4787]: I1203 19:15:30.343164 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:15:30 crc kubenswrapper[4787]: E1203 19:15:30.460465 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice/crio-17f8cd8c2cc7e5db5c49cbf7663cc452ac345066258872b63223da6f82163ffe\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice\": RecentStats: unable to find data in memory cache]" Dec 03 19:15:30 crc kubenswrapper[4787]: I1203 19:15:30.749579 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"dba0eca5-4008-4ce5-b555-d6db725f6466","Type":"ContainerStarted","Data":"6827790696bf4e10023f202eb462ce7115c3fc12397d20fb37d89ffe03f18ea6"} Dec 03 19:15:32 crc kubenswrapper[4787]: I1203 19:15:32.335994 4787 scope.go:117] "RemoveContainer" containerID="c760cab848a34b7412baaa1bbdc8c4673ad19efda3d72a618b3a8411de801ac0" Dec 03 19:15:32 crc kubenswrapper[4787]: I1203 19:15:32.778533 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"dba0eca5-4008-4ce5-b555-d6db725f6466","Type":"ContainerStarted","Data":"cdcb96d463b1d2ce7004a9f2eea9d5ce962752ce32adca892437a07d7d52429e"} Dec 03 19:15:32 crc kubenswrapper[4787]: I1203 19:15:32.806949 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=2.011058492 podStartE2EDuration="3.806902077s" podCreationTimestamp="2025-12-03 19:15:29 +0000 UTC" firstStartedPulling="2025-12-03 19:15:30.342844047 +0000 UTC m=+7347.160315006" lastFinishedPulling="2025-12-03 19:15:32.138687632 +0000 UTC m=+7348.956158591" observedRunningTime="2025-12-03 19:15:32.793776682 +0000 UTC m=+7349.611247651" watchObservedRunningTime="2025-12-03 19:15:32.806902077 +0000 UTC m=+7349.624373046" Dec 03 19:15:40 crc kubenswrapper[4787]: E1203 19:15:40.760414 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice/crio-17f8cd8c2cc7e5db5c49cbf7663cc452ac345066258872b63223da6f82163ffe\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice\": RecentStats: unable to find data in memory cache]" Dec 03 19:15:51 crc kubenswrapper[4787]: E1203 19:15:51.035476 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice/crio-17f8cd8c2cc7e5db5c49cbf7663cc452ac345066258872b63223da6f82163ffe\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice\": RecentStats: unable to find data in memory cache]" Dec 03 19:16:01 crc kubenswrapper[4787]: E1203 19:16:01.395139 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice/crio-17f8cd8c2cc7e5db5c49cbf7663cc452ac345066258872b63223da6f82163ffe\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode38b6288_b87a_4bc4_913b_368f2c86739a.slice\": RecentStats: unable to find data in memory cache]" Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.422881 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vj7r9/must-gather-l8k5x"] Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.425656 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/must-gather-l8k5x" Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.429781 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vj7r9"/"kube-root-ca.crt" Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.429874 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-vj7r9"/"default-dockercfg-7pldz" Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.429793 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vj7r9"/"openshift-service-ca.crt" Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.489852 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvrtk\" (UniqueName: \"kubernetes.io/projected/a06ae924-3061-4b7e-82fb-345f03981a85-kube-api-access-qvrtk\") pod \"must-gather-l8k5x\" (UID: \"a06ae924-3061-4b7e-82fb-345f03981a85\") " pod="openshift-must-gather-vj7r9/must-gather-l8k5x" Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.490146 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a06ae924-3061-4b7e-82fb-345f03981a85-must-gather-output\") pod \"must-gather-l8k5x\" (UID: \"a06ae924-3061-4b7e-82fb-345f03981a85\") " pod="openshift-must-gather-vj7r9/must-gather-l8k5x" Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.506162 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vj7r9/must-gather-l8k5x"] Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.591623 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a06ae924-3061-4b7e-82fb-345f03981a85-must-gather-output\") pod \"must-gather-l8k5x\" (UID: \"a06ae924-3061-4b7e-82fb-345f03981a85\") " pod="openshift-must-gather-vj7r9/must-gather-l8k5x" Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.591741 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvrtk\" (UniqueName: \"kubernetes.io/projected/a06ae924-3061-4b7e-82fb-345f03981a85-kube-api-access-qvrtk\") pod \"must-gather-l8k5x\" (UID: \"a06ae924-3061-4b7e-82fb-345f03981a85\") " pod="openshift-must-gather-vj7r9/must-gather-l8k5x" Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.592299 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a06ae924-3061-4b7e-82fb-345f03981a85-must-gather-output\") pod \"must-gather-l8k5x\" (UID: \"a06ae924-3061-4b7e-82fb-345f03981a85\") " pod="openshift-must-gather-vj7r9/must-gather-l8k5x" Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.616982 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvrtk\" (UniqueName: \"kubernetes.io/projected/a06ae924-3061-4b7e-82fb-345f03981a85-kube-api-access-qvrtk\") pod \"must-gather-l8k5x\" (UID: \"a06ae924-3061-4b7e-82fb-345f03981a85\") " pod="openshift-must-gather-vj7r9/must-gather-l8k5x" Dec 03 19:16:01 crc kubenswrapper[4787]: I1203 19:16:01.749448 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/must-gather-l8k5x" Dec 03 19:16:02 crc kubenswrapper[4787]: W1203 19:16:02.438594 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda06ae924_3061_4b7e_82fb_345f03981a85.slice/crio-ea0cc4b638b4cf0c047519ff7f072a436879e53962b0195516fbcfbc07cd24bb WatchSource:0}: Error finding container ea0cc4b638b4cf0c047519ff7f072a436879e53962b0195516fbcfbc07cd24bb: Status 404 returned error can't find the container with id ea0cc4b638b4cf0c047519ff7f072a436879e53962b0195516fbcfbc07cd24bb Dec 03 19:16:02 crc kubenswrapper[4787]: I1203 19:16:02.440724 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vj7r9/must-gather-l8k5x"] Dec 03 19:16:03 crc kubenswrapper[4787]: I1203 19:16:03.165581 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/must-gather-l8k5x" event={"ID":"a06ae924-3061-4b7e-82fb-345f03981a85","Type":"ContainerStarted","Data":"ea0cc4b638b4cf0c047519ff7f072a436879e53962b0195516fbcfbc07cd24bb"} Dec 03 19:16:10 crc kubenswrapper[4787]: I1203 19:16:10.256052 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/must-gather-l8k5x" event={"ID":"a06ae924-3061-4b7e-82fb-345f03981a85","Type":"ContainerStarted","Data":"5951d2705c1baed53643425dcb9418057aa1bbe056abf41f4d5a80bcdf11e4ba"} Dec 03 19:16:10 crc kubenswrapper[4787]: I1203 19:16:10.256555 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/must-gather-l8k5x" event={"ID":"a06ae924-3061-4b7e-82fb-345f03981a85","Type":"ContainerStarted","Data":"4d9573ecfa71b743ca24a7534b5a225e65b6bb72bab6df25e461a107813957f4"} Dec 03 19:16:10 crc kubenswrapper[4787]: I1203 19:16:10.270525 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vj7r9/must-gather-l8k5x" podStartSLOduration=2.628638492 podStartE2EDuration="9.270496408s" podCreationTimestamp="2025-12-03 19:16:01 +0000 UTC" firstStartedPulling="2025-12-03 19:16:02.447154137 +0000 UTC m=+7379.264625126" lastFinishedPulling="2025-12-03 19:16:09.089012073 +0000 UTC m=+7385.906483042" observedRunningTime="2025-12-03 19:16:10.268704421 +0000 UTC m=+7387.086175440" watchObservedRunningTime="2025-12-03 19:16:10.270496408 +0000 UTC m=+7387.087967407" Dec 03 19:16:14 crc kubenswrapper[4787]: I1203 19:16:14.801032 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vj7r9/crc-debug-twx2p"] Dec 03 19:16:14 crc kubenswrapper[4787]: I1203 19:16:14.803229 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-twx2p" Dec 03 19:16:14 crc kubenswrapper[4787]: I1203 19:16:14.838475 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8d8q\" (UniqueName: \"kubernetes.io/projected/36d234e6-613c-4745-9e0d-6d4bb3f17162-kube-api-access-j8d8q\") pod \"crc-debug-twx2p\" (UID: \"36d234e6-613c-4745-9e0d-6d4bb3f17162\") " pod="openshift-must-gather-vj7r9/crc-debug-twx2p" Dec 03 19:16:14 crc kubenswrapper[4787]: I1203 19:16:14.838525 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36d234e6-613c-4745-9e0d-6d4bb3f17162-host\") pod \"crc-debug-twx2p\" (UID: \"36d234e6-613c-4745-9e0d-6d4bb3f17162\") " pod="openshift-must-gather-vj7r9/crc-debug-twx2p" Dec 03 19:16:14 crc kubenswrapper[4787]: I1203 19:16:14.941252 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8d8q\" (UniqueName: \"kubernetes.io/projected/36d234e6-613c-4745-9e0d-6d4bb3f17162-kube-api-access-j8d8q\") pod \"crc-debug-twx2p\" (UID: \"36d234e6-613c-4745-9e0d-6d4bb3f17162\") " pod="openshift-must-gather-vj7r9/crc-debug-twx2p" Dec 03 19:16:14 crc kubenswrapper[4787]: I1203 19:16:14.941303 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36d234e6-613c-4745-9e0d-6d4bb3f17162-host\") pod \"crc-debug-twx2p\" (UID: \"36d234e6-613c-4745-9e0d-6d4bb3f17162\") " pod="openshift-must-gather-vj7r9/crc-debug-twx2p" Dec 03 19:16:14 crc kubenswrapper[4787]: I1203 19:16:14.941516 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36d234e6-613c-4745-9e0d-6d4bb3f17162-host\") pod \"crc-debug-twx2p\" (UID: \"36d234e6-613c-4745-9e0d-6d4bb3f17162\") " pod="openshift-must-gather-vj7r9/crc-debug-twx2p" Dec 03 19:16:14 crc kubenswrapper[4787]: I1203 19:16:14.964926 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8d8q\" (UniqueName: \"kubernetes.io/projected/36d234e6-613c-4745-9e0d-6d4bb3f17162-kube-api-access-j8d8q\") pod \"crc-debug-twx2p\" (UID: \"36d234e6-613c-4745-9e0d-6d4bb3f17162\") " pod="openshift-must-gather-vj7r9/crc-debug-twx2p" Dec 03 19:16:15 crc kubenswrapper[4787]: I1203 19:16:15.120087 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-twx2p" Dec 03 19:16:15 crc kubenswrapper[4787]: I1203 19:16:15.313085 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/crc-debug-twx2p" event={"ID":"36d234e6-613c-4745-9e0d-6d4bb3f17162","Type":"ContainerStarted","Data":"d269538a584bd13fc8caa39f330a363439ad54e50da0f445e9783eccfe33da79"} Dec 03 19:16:18 crc kubenswrapper[4787]: I1203 19:16:18.990041 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:16:18 crc kubenswrapper[4787]: I1203 19:16:18.990571 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.439251 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-gm2cz"] Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.443597 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.465545 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gm2cz"] Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.575873 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-utilities\") pod \"redhat-operators-gm2cz\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.576373 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4bqfc\" (UniqueName: \"kubernetes.io/projected/56d71e9f-aa7f-4030-961c-b279aeff073d-kube-api-access-4bqfc\") pod \"redhat-operators-gm2cz\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.576525 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-catalog-content\") pod \"redhat-operators-gm2cz\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.678355 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4bqfc\" (UniqueName: \"kubernetes.io/projected/56d71e9f-aa7f-4030-961c-b279aeff073d-kube-api-access-4bqfc\") pod \"redhat-operators-gm2cz\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.678646 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-catalog-content\") pod \"redhat-operators-gm2cz\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.678710 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-utilities\") pod \"redhat-operators-gm2cz\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.679204 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-catalog-content\") pod \"redhat-operators-gm2cz\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.679223 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-utilities\") pod \"redhat-operators-gm2cz\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.699041 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4bqfc\" (UniqueName: \"kubernetes.io/projected/56d71e9f-aa7f-4030-961c-b279aeff073d-kube-api-access-4bqfc\") pod \"redhat-operators-gm2cz\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:20 crc kubenswrapper[4787]: I1203 19:16:20.778319 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:21 crc kubenswrapper[4787]: I1203 19:16:21.444916 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-gm2cz"] Dec 03 19:16:21 crc kubenswrapper[4787]: I1203 19:16:21.581414 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gm2cz" event={"ID":"56d71e9f-aa7f-4030-961c-b279aeff073d","Type":"ContainerStarted","Data":"efdc7836cfd29b04412246392461f1b48824a6177203415146fb5e91623173ce"} Dec 03 19:16:22 crc kubenswrapper[4787]: I1203 19:16:22.595454 4787 generic.go:334] "Generic (PLEG): container finished" podID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerID="a0f7551107bb961bae625c7b84a75715b46b76febfce0225468f00dcecbdf5de" exitCode=0 Dec 03 19:16:22 crc kubenswrapper[4787]: I1203 19:16:22.595848 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gm2cz" event={"ID":"56d71e9f-aa7f-4030-961c-b279aeff073d","Type":"ContainerDied","Data":"a0f7551107bb961bae625c7b84a75715b46b76febfce0225468f00dcecbdf5de"} Dec 03 19:16:25 crc kubenswrapper[4787]: I1203 19:16:25.631551 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gm2cz" event={"ID":"56d71e9f-aa7f-4030-961c-b279aeff073d","Type":"ContainerStarted","Data":"95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569"} Dec 03 19:16:28 crc kubenswrapper[4787]: I1203 19:16:28.780312 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="metallb-system/frr-k8s-pgt2s" podUID="37601185-c3bd-4614-b119-05f1b07f2875" containerName="frr" probeResult="failure" output="Get \"http://127.0.0.1:7573/livez\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 19:16:28 crc kubenswrapper[4787]: I1203 19:16:28.789274 4787 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.859453047s: [/var/lib/containers/storage/overlay/eac864960c64f584b720d7322fa2c7951cd97b4b0ff2ef74f991b9799f0c9650/diff /var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-bzjxn_43962cfc-342c-49db-83bf-ccde92708a0b/kube-rbac-proxy/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 19:16:28 crc kubenswrapper[4787]: I1203 19:16:28.789452 4787 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.874577215s: [/var/lib/containers/storage/overlay/a0af484c6b1298cad48ca4018a4b58405f57f0ad3405c7ff59d57ae8ff37add0/diff /var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-fgdm7_6d12a43d-d3da-4b99-b48b-519d660d2527/kube-rbac-proxy/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 19:16:28 crc kubenswrapper[4787]: I1203 19:16:28.804338 4787 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.87631288s: [/var/lib/containers/storage/overlay/fbda168b6cc5cf99c66f39e0cb0768965f5420f493590d9185292fa0107a528b/diff /var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-75r2j_67472bf9-12b8-4463-bc95-dec19e689f36/kube-rbac-proxy/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 19:16:28 crc kubenswrapper[4787]: I1203 19:16:28.809476 4787 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.186747085s: [/var/lib/containers/storage/overlay/b46b60372886f7bff1062f41a2542e54f2b90837e4c092bcd969a9a3cece762c/diff /var/log/pods/openshift-logging_logging-loki-gateway-57f76f4c5d-v45xd_cb532cad-464c-4fe6-99de-b7897de9ac51/gateway/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 19:16:29 crc kubenswrapper[4787]: I1203 19:16:29.478343 4787 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 2.331379s: [/var/lib/containers/storage/overlay/6ba4c9051d70e3e3eb09b01fe02158e962831aa3da91614616a00c68f21a8ed0/diff /var/log/pods/openstack_keystone-69bdc498c7-wqxdv_c244629d-4ad6-4a6f-9f3d-eaceda01c7e8/keystone-api/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 19:16:29 crc kubenswrapper[4787]: I1203 19:16:29.479198 4787 trace.go:236] Trace[1939571411]: "Calculate volume metrics of registry-storage for pod openshift-image-registry/image-registry-66df7c8f76-nr9fq" (03-Dec-2025 19:16:27.819) (total time: 1659ms): Dec 03 19:16:29 crc kubenswrapper[4787]: Trace[1939571411]: [1.659086968s] [1.659086968s] END Dec 03 19:16:36 crc kubenswrapper[4787]: I1203 19:16:36.774579 4787 generic.go:334] "Generic (PLEG): container finished" podID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerID="95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569" exitCode=0 Dec 03 19:16:36 crc kubenswrapper[4787]: I1203 19:16:36.774663 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gm2cz" event={"ID":"56d71e9f-aa7f-4030-961c-b279aeff073d","Type":"ContainerDied","Data":"95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569"} Dec 03 19:16:42 crc kubenswrapper[4787]: I1203 19:16:42.300348 4787 patch_prober.go:28] interesting pod/router-default-5444994796-x2wsr container/router namespace/openshift-ingress: Liveness probe status=failure output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 19:16:42 crc kubenswrapper[4787]: I1203 19:16:42.301386 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-ingress/router-default-5444994796-x2wsr" podUID="70235f57-caf7-48d6-ab8e-85230e423cd0" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 19:16:42 crc kubenswrapper[4787]: I1203 19:16:42.300365 4787 patch_prober.go:28] interesting pod/router-default-5444994796-x2wsr container/router namespace/openshift-ingress: Readiness probe status=failure output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 19:16:42 crc kubenswrapper[4787]: I1203 19:16:42.301674 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-ingress/router-default-5444994796-x2wsr" podUID="70235f57-caf7-48d6-ab8e-85230e423cd0" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 19:16:42 crc kubenswrapper[4787]: I1203 19:16:42.747167 4787 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-gwb62 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Liveness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 19:16:42 crc kubenswrapper[4787]: I1203 19:16:42.747217 4787 patch_prober.go:28] interesting pod/package-server-manager-789f6589d5-gwb62 container/package-server-manager namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 03 19:16:42 crc kubenswrapper[4787]: I1203 19:16:42.747299 4787 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" podUID="97cab125-72b2-454c-9bef-bd70f5eb9654" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 19:16:42 crc kubenswrapper[4787]: I1203 19:16:42.747220 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-gwb62" podUID="97cab125-72b2-454c-9bef-bd70f5eb9654" containerName="package-server-manager" probeResult="failure" output="Get \"http://10.217.0.35:8080/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 03 19:16:42 crc kubenswrapper[4787]: I1203 19:16:42.771514 4787 trace.go:236] Trace[157786993]: "Calculate volume metrics of config-data-generated for pod openstack/openstack-galera-0" (03-Dec-2025 19:16:41.482) (total time: 1289ms): Dec 03 19:16:42 crc kubenswrapper[4787]: Trace[157786993]: [1.289304682s] [1.289304682s] END Dec 03 19:16:42 crc kubenswrapper[4787]: I1203 19:16:42.773258 4787 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.506867273s: [/var/lib/containers/storage/overlay/2bd301c5bdcc99b106bad424d7c51f887cb2614794f1b81bd27d4a4a577b7350/diff /var/log/pods/openstack_barbican-keystone-listener-78b7c4ccf4-bxfhc_d83ca165-acb3-4e27-b8f3-519897743134/barbican-keystone-listener/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 19:16:42 crc kubenswrapper[4787]: I1203 19:16:42.774321 4787 fsHandler.go:133] fs: disk usage and inodes count on following dirs took 1.729493829s: [/var/lib/containers/storage/overlay/df50d677b5034f00ff66954b4946b80dc8e118ec22130a79d9c7e3f76cb621e5/diff /var/log/pods/openstack_barbican-worker-5f6df7d97f-5vmdt_65180fdc-f51b-45e7-ac70-05b9489e1201/barbican-worker/0.log]; will not log again for this container unless duration exceeds 2s Dec 03 19:16:44 crc kubenswrapper[4787]: E1203 19:16:44.570152 4787 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Dec 03 19:16:44 crc kubenswrapper[4787]: E1203 19:16:44.571593 4787 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-j8d8q,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-twx2p_openshift-must-gather-vj7r9(36d234e6-613c-4745-9e0d-6d4bb3f17162): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 03 19:16:44 crc kubenswrapper[4787]: E1203 19:16:44.572803 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-vj7r9/crc-debug-twx2p" podUID="36d234e6-613c-4745-9e0d-6d4bb3f17162" Dec 03 19:16:44 crc kubenswrapper[4787]: E1203 19:16:44.906552 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-vj7r9/crc-debug-twx2p" podUID="36d234e6-613c-4745-9e0d-6d4bb3f17162" Dec 03 19:16:45 crc kubenswrapper[4787]: I1203 19:16:45.913753 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gm2cz" event={"ID":"56d71e9f-aa7f-4030-961c-b279aeff073d","Type":"ContainerStarted","Data":"e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d"} Dec 03 19:16:45 crc kubenswrapper[4787]: I1203 19:16:45.934210 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-gm2cz" podStartSLOduration=3.661823155 podStartE2EDuration="25.934191826s" podCreationTimestamp="2025-12-03 19:16:20 +0000 UTC" firstStartedPulling="2025-12-03 19:16:22.59859915 +0000 UTC m=+7399.416070109" lastFinishedPulling="2025-12-03 19:16:44.870967821 +0000 UTC m=+7421.688438780" observedRunningTime="2025-12-03 19:16:45.931995048 +0000 UTC m=+7422.749465997" watchObservedRunningTime="2025-12-03 19:16:45.934191826 +0000 UTC m=+7422.751662775" Dec 03 19:16:48 crc kubenswrapper[4787]: I1203 19:16:48.989524 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:16:48 crc kubenswrapper[4787]: I1203 19:16:48.989867 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:16:50 crc kubenswrapper[4787]: I1203 19:16:50.779519 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:50 crc kubenswrapper[4787]: I1203 19:16:50.781532 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:16:51 crc kubenswrapper[4787]: I1203 19:16:51.841913 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gm2cz" podUID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerName="registry-server" probeResult="failure" output=< Dec 03 19:16:51 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 19:16:51 crc kubenswrapper[4787]: > Dec 03 19:17:01 crc kubenswrapper[4787]: I1203 19:17:01.103969 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/crc-debug-twx2p" event={"ID":"36d234e6-613c-4745-9e0d-6d4bb3f17162","Type":"ContainerStarted","Data":"85917550bd18b5a593b6b563a5367889774efa3ccba9c571784577eb0b4b2e44"} Dec 03 19:17:01 crc kubenswrapper[4787]: I1203 19:17:01.131602 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vj7r9/crc-debug-twx2p" podStartSLOduration=1.704464911 podStartE2EDuration="47.131584158s" podCreationTimestamp="2025-12-03 19:16:14 +0000 UTC" firstStartedPulling="2025-12-03 19:16:15.16567721 +0000 UTC m=+7391.983148179" lastFinishedPulling="2025-12-03 19:17:00.592796447 +0000 UTC m=+7437.410267426" observedRunningTime="2025-12-03 19:17:01.119158041 +0000 UTC m=+7437.936629000" watchObservedRunningTime="2025-12-03 19:17:01.131584158 +0000 UTC m=+7437.949055117" Dec 03 19:17:01 crc kubenswrapper[4787]: I1203 19:17:01.850604 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-gm2cz" podUID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerName="registry-server" probeResult="failure" output=< Dec 03 19:17:01 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 19:17:01 crc kubenswrapper[4787]: > Dec 03 19:17:10 crc kubenswrapper[4787]: I1203 19:17:10.838390 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:17:10 crc kubenswrapper[4787]: I1203 19:17:10.899473 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:17:11 crc kubenswrapper[4787]: I1203 19:17:11.077292 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gm2cz"] Dec 03 19:17:12 crc kubenswrapper[4787]: I1203 19:17:12.228345 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-gm2cz" podUID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerName="registry-server" containerID="cri-o://e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d" gracePeriod=2 Dec 03 19:17:12 crc kubenswrapper[4787]: I1203 19:17:12.882545 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.009730 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-utilities\") pod \"56d71e9f-aa7f-4030-961c-b279aeff073d\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.009797 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-catalog-content\") pod \"56d71e9f-aa7f-4030-961c-b279aeff073d\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.009825 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4bqfc\" (UniqueName: \"kubernetes.io/projected/56d71e9f-aa7f-4030-961c-b279aeff073d-kube-api-access-4bqfc\") pod \"56d71e9f-aa7f-4030-961c-b279aeff073d\" (UID: \"56d71e9f-aa7f-4030-961c-b279aeff073d\") " Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.012049 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-utilities" (OuterVolumeSpecName: "utilities") pod "56d71e9f-aa7f-4030-961c-b279aeff073d" (UID: "56d71e9f-aa7f-4030-961c-b279aeff073d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.028291 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/56d71e9f-aa7f-4030-961c-b279aeff073d-kube-api-access-4bqfc" (OuterVolumeSpecName: "kube-api-access-4bqfc") pod "56d71e9f-aa7f-4030-961c-b279aeff073d" (UID: "56d71e9f-aa7f-4030-961c-b279aeff073d"). InnerVolumeSpecName "kube-api-access-4bqfc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.112661 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.112707 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4bqfc\" (UniqueName: \"kubernetes.io/projected/56d71e9f-aa7f-4030-961c-b279aeff073d-kube-api-access-4bqfc\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.130486 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "56d71e9f-aa7f-4030-961c-b279aeff073d" (UID: "56d71e9f-aa7f-4030-961c-b279aeff073d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.215263 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/56d71e9f-aa7f-4030-961c-b279aeff073d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.240243 4787 generic.go:334] "Generic (PLEG): container finished" podID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerID="e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d" exitCode=0 Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.240315 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-gm2cz" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.240306 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gm2cz" event={"ID":"56d71e9f-aa7f-4030-961c-b279aeff073d","Type":"ContainerDied","Data":"e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d"} Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.240382 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-gm2cz" event={"ID":"56d71e9f-aa7f-4030-961c-b279aeff073d","Type":"ContainerDied","Data":"efdc7836cfd29b04412246392461f1b48824a6177203415146fb5e91623173ce"} Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.240409 4787 scope.go:117] "RemoveContainer" containerID="e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.279436 4787 scope.go:117] "RemoveContainer" containerID="95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.283540 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-gm2cz"] Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.294351 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-gm2cz"] Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.301204 4787 scope.go:117] "RemoveContainer" containerID="a0f7551107bb961bae625c7b84a75715b46b76febfce0225468f00dcecbdf5de" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.367066 4787 scope.go:117] "RemoveContainer" containerID="e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d" Dec 03 19:17:13 crc kubenswrapper[4787]: E1203 19:17:13.367553 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d\": container with ID starting with e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d not found: ID does not exist" containerID="e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.367665 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d"} err="failed to get container status \"e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d\": rpc error: code = NotFound desc = could not find container \"e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d\": container with ID starting with e776cbbc9290372884eabffab74a8786144c9130728a865a7a088c7844cefc0d not found: ID does not exist" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.367750 4787 scope.go:117] "RemoveContainer" containerID="95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569" Dec 03 19:17:13 crc kubenswrapper[4787]: E1203 19:17:13.367992 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569\": container with ID starting with 95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569 not found: ID does not exist" containerID="95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.368081 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569"} err="failed to get container status \"95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569\": rpc error: code = NotFound desc = could not find container \"95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569\": container with ID starting with 95987a6b787f76e6a6f3d202c4d5be2a060fea02ba663a8811c9bc3f7a8ed569 not found: ID does not exist" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.368143 4787 scope.go:117] "RemoveContainer" containerID="a0f7551107bb961bae625c7b84a75715b46b76febfce0225468f00dcecbdf5de" Dec 03 19:17:13 crc kubenswrapper[4787]: E1203 19:17:13.369110 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0f7551107bb961bae625c7b84a75715b46b76febfce0225468f00dcecbdf5de\": container with ID starting with a0f7551107bb961bae625c7b84a75715b46b76febfce0225468f00dcecbdf5de not found: ID does not exist" containerID="a0f7551107bb961bae625c7b84a75715b46b76febfce0225468f00dcecbdf5de" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.369187 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0f7551107bb961bae625c7b84a75715b46b76febfce0225468f00dcecbdf5de"} err="failed to get container status \"a0f7551107bb961bae625c7b84a75715b46b76febfce0225468f00dcecbdf5de\": rpc error: code = NotFound desc = could not find container \"a0f7551107bb961bae625c7b84a75715b46b76febfce0225468f00dcecbdf5de\": container with ID starting with a0f7551107bb961bae625c7b84a75715b46b76febfce0225468f00dcecbdf5de not found: ID does not exist" Dec 03 19:17:13 crc kubenswrapper[4787]: I1203 19:17:13.779882 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="56d71e9f-aa7f-4030-961c-b279aeff073d" path="/var/lib/kubelet/pods/56d71e9f-aa7f-4030-961c-b279aeff073d/volumes" Dec 03 19:17:18 crc kubenswrapper[4787]: I1203 19:17:18.990125 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:17:18 crc kubenswrapper[4787]: I1203 19:17:18.990645 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:17:18 crc kubenswrapper[4787]: I1203 19:17:18.990683 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 19:17:18 crc kubenswrapper[4787]: I1203 19:17:18.991467 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:17:18 crc kubenswrapper[4787]: I1203 19:17:18.991508 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" gracePeriod=600 Dec 03 19:17:19 crc kubenswrapper[4787]: E1203 19:17:19.144061 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:17:19 crc kubenswrapper[4787]: I1203 19:17:19.308711 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" exitCode=0 Dec 03 19:17:19 crc kubenswrapper[4787]: I1203 19:17:19.308755 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514"} Dec 03 19:17:19 crc kubenswrapper[4787]: I1203 19:17:19.308793 4787 scope.go:117] "RemoveContainer" containerID="65214d10d37779bf02fb7aef42e332ae9add4645b6632d81a6d49b8d359c7917" Dec 03 19:17:19 crc kubenswrapper[4787]: I1203 19:17:19.311365 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:17:19 crc kubenswrapper[4787]: E1203 19:17:19.311921 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.681618 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nz8q4"] Dec 03 19:17:21 crc kubenswrapper[4787]: E1203 19:17:21.682685 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerName="extract-utilities" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.682697 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerName="extract-utilities" Dec 03 19:17:21 crc kubenswrapper[4787]: E1203 19:17:21.682737 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerName="extract-content" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.682746 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerName="extract-content" Dec 03 19:17:21 crc kubenswrapper[4787]: E1203 19:17:21.682769 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerName="registry-server" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.682776 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerName="registry-server" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.683007 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="56d71e9f-aa7f-4030-961c-b279aeff073d" containerName="registry-server" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.684667 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.698495 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nz8q4"] Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.767995 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-utilities\") pod \"community-operators-nz8q4\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.768388 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bz8hd\" (UniqueName: \"kubernetes.io/projected/d737f2b9-b321-4104-967d-415f3c9c72f1-kube-api-access-bz8hd\") pod \"community-operators-nz8q4\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.768429 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-catalog-content\") pod \"community-operators-nz8q4\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.869923 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-nd2pw"] Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.870784 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bz8hd\" (UniqueName: \"kubernetes.io/projected/d737f2b9-b321-4104-967d-415f3c9c72f1-kube-api-access-bz8hd\") pod \"community-operators-nz8q4\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.870857 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-catalog-content\") pod \"community-operators-nz8q4\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.871038 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-utilities\") pod \"community-operators-nz8q4\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.871517 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-utilities\") pod \"community-operators-nz8q4\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.872455 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-catalog-content\") pod \"community-operators-nz8q4\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.873637 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.883746 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nd2pw"] Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.917952 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bz8hd\" (UniqueName: \"kubernetes.io/projected/d737f2b9-b321-4104-967d-415f3c9c72f1-kube-api-access-bz8hd\") pod \"community-operators-nz8q4\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.973280 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5xwjz\" (UniqueName: \"kubernetes.io/projected/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-kube-api-access-5xwjz\") pod \"certified-operators-nd2pw\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.973604 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-catalog-content\") pod \"certified-operators-nd2pw\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:21 crc kubenswrapper[4787]: I1203 19:17:21.973824 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-utilities\") pod \"certified-operators-nd2pw\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:22 crc kubenswrapper[4787]: I1203 19:17:22.025972 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:22 crc kubenswrapper[4787]: I1203 19:17:22.076429 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-utilities\") pod \"certified-operators-nd2pw\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:22 crc kubenswrapper[4787]: I1203 19:17:22.076719 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5xwjz\" (UniqueName: \"kubernetes.io/projected/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-kube-api-access-5xwjz\") pod \"certified-operators-nd2pw\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:22 crc kubenswrapper[4787]: I1203 19:17:22.076877 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-catalog-content\") pod \"certified-operators-nd2pw\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:22 crc kubenswrapper[4787]: I1203 19:17:22.076966 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-utilities\") pod \"certified-operators-nd2pw\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:22 crc kubenswrapper[4787]: I1203 19:17:22.077511 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-catalog-content\") pod \"certified-operators-nd2pw\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:22 crc kubenswrapper[4787]: I1203 19:17:22.099784 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5xwjz\" (UniqueName: \"kubernetes.io/projected/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-kube-api-access-5xwjz\") pod \"certified-operators-nd2pw\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:22 crc kubenswrapper[4787]: I1203 19:17:22.195451 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:22 crc kubenswrapper[4787]: I1203 19:17:22.521193 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nz8q4"] Dec 03 19:17:22 crc kubenswrapper[4787]: I1203 19:17:22.856839 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-nd2pw"] Dec 03 19:17:22 crc kubenswrapper[4787]: W1203 19:17:22.862058 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9395c2b0_67a7_4a8d_a89a_026dc50f4ae2.slice/crio-18f6ada27532f44902d48d9cd1a138e4e189b7364536b8c37ea88ed69d441d6b WatchSource:0}: Error finding container 18f6ada27532f44902d48d9cd1a138e4e189b7364536b8c37ea88ed69d441d6b: Status 404 returned error can't find the container with id 18f6ada27532f44902d48d9cd1a138e4e189b7364536b8c37ea88ed69d441d6b Dec 03 19:17:23 crc kubenswrapper[4787]: I1203 19:17:23.358303 4787 generic.go:334] "Generic (PLEG): container finished" podID="d737f2b9-b321-4104-967d-415f3c9c72f1" containerID="980f85f85cf89934ad5ed8093919a3baaa0afb7df6b23cb4bb83c3a690b889ae" exitCode=0 Dec 03 19:17:23 crc kubenswrapper[4787]: I1203 19:17:23.358399 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz8q4" event={"ID":"d737f2b9-b321-4104-967d-415f3c9c72f1","Type":"ContainerDied","Data":"980f85f85cf89934ad5ed8093919a3baaa0afb7df6b23cb4bb83c3a690b889ae"} Dec 03 19:17:23 crc kubenswrapper[4787]: I1203 19:17:23.358591 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz8q4" event={"ID":"d737f2b9-b321-4104-967d-415f3c9c72f1","Type":"ContainerStarted","Data":"cfea4fe240422914b3aa1570594317bc5e399543ba4e9d9085c9314deeb82df9"} Dec 03 19:17:23 crc kubenswrapper[4787]: I1203 19:17:23.360478 4787 generic.go:334] "Generic (PLEG): container finished" podID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" containerID="3e9a57d146ec736cde022b223b9bebe29b241303deef9cedf8e6634726812fea" exitCode=0 Dec 03 19:17:23 crc kubenswrapper[4787]: I1203 19:17:23.360502 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd2pw" event={"ID":"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2","Type":"ContainerDied","Data":"3e9a57d146ec736cde022b223b9bebe29b241303deef9cedf8e6634726812fea"} Dec 03 19:17:23 crc kubenswrapper[4787]: I1203 19:17:23.360517 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd2pw" event={"ID":"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2","Type":"ContainerStarted","Data":"18f6ada27532f44902d48d9cd1a138e4e189b7364536b8c37ea88ed69d441d6b"} Dec 03 19:17:24 crc kubenswrapper[4787]: I1203 19:17:24.373705 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd2pw" event={"ID":"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2","Type":"ContainerStarted","Data":"52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708"} Dec 03 19:17:24 crc kubenswrapper[4787]: I1203 19:17:24.376856 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz8q4" event={"ID":"d737f2b9-b321-4104-967d-415f3c9c72f1","Type":"ContainerStarted","Data":"7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a"} Dec 03 19:17:27 crc kubenswrapper[4787]: I1203 19:17:27.407987 4787 generic.go:334] "Generic (PLEG): container finished" podID="d737f2b9-b321-4104-967d-415f3c9c72f1" containerID="7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a" exitCode=0 Dec 03 19:17:27 crc kubenswrapper[4787]: I1203 19:17:27.408173 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz8q4" event={"ID":"d737f2b9-b321-4104-967d-415f3c9c72f1","Type":"ContainerDied","Data":"7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a"} Dec 03 19:17:27 crc kubenswrapper[4787]: I1203 19:17:27.411444 4787 generic.go:334] "Generic (PLEG): container finished" podID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" containerID="52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708" exitCode=0 Dec 03 19:17:27 crc kubenswrapper[4787]: I1203 19:17:27.411472 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd2pw" event={"ID":"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2","Type":"ContainerDied","Data":"52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708"} Dec 03 19:17:28 crc kubenswrapper[4787]: I1203 19:17:28.438992 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd2pw" event={"ID":"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2","Type":"ContainerStarted","Data":"9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900"} Dec 03 19:17:28 crc kubenswrapper[4787]: I1203 19:17:28.447625 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz8q4" event={"ID":"d737f2b9-b321-4104-967d-415f3c9c72f1","Type":"ContainerStarted","Data":"4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e"} Dec 03 19:17:28 crc kubenswrapper[4787]: I1203 19:17:28.472109 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-nd2pw" podStartSLOduration=2.990042233 podStartE2EDuration="7.472089741s" podCreationTimestamp="2025-12-03 19:17:21 +0000 UTC" firstStartedPulling="2025-12-03 19:17:23.362642252 +0000 UTC m=+7460.180113211" lastFinishedPulling="2025-12-03 19:17:27.84468976 +0000 UTC m=+7464.662160719" observedRunningTime="2025-12-03 19:17:28.458555185 +0000 UTC m=+7465.276026174" watchObservedRunningTime="2025-12-03 19:17:28.472089741 +0000 UTC m=+7465.289560700" Dec 03 19:17:28 crc kubenswrapper[4787]: I1203 19:17:28.492117 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nz8q4" podStartSLOduration=2.985413632 podStartE2EDuration="7.492092497s" podCreationTimestamp="2025-12-03 19:17:21 +0000 UTC" firstStartedPulling="2025-12-03 19:17:23.360094895 +0000 UTC m=+7460.177565854" lastFinishedPulling="2025-12-03 19:17:27.86677376 +0000 UTC m=+7464.684244719" observedRunningTime="2025-12-03 19:17:28.489106158 +0000 UTC m=+7465.306577127" watchObservedRunningTime="2025-12-03 19:17:28.492092497 +0000 UTC m=+7465.309563466" Dec 03 19:17:30 crc kubenswrapper[4787]: I1203 19:17:30.766700 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:17:30 crc kubenswrapper[4787]: E1203 19:17:30.767293 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:17:32 crc kubenswrapper[4787]: I1203 19:17:32.026901 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:32 crc kubenswrapper[4787]: I1203 19:17:32.027268 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:32 crc kubenswrapper[4787]: I1203 19:17:32.083819 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:32 crc kubenswrapper[4787]: I1203 19:17:32.196171 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:32 crc kubenswrapper[4787]: I1203 19:17:32.196613 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:32 crc kubenswrapper[4787]: I1203 19:17:32.269241 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:33 crc kubenswrapper[4787]: I1203 19:17:33.549981 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:35 crc kubenswrapper[4787]: I1203 19:17:35.866192 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nd2pw"] Dec 03 19:17:35 crc kubenswrapper[4787]: I1203 19:17:35.866934 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-nd2pw" podUID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" containerName="registry-server" containerID="cri-o://9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900" gracePeriod=2 Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.403755 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.456246 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5xwjz\" (UniqueName: \"kubernetes.io/projected/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-kube-api-access-5xwjz\") pod \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.456346 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-utilities\") pod \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.456639 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-catalog-content\") pod \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\" (UID: \"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2\") " Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.459418 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-utilities" (OuterVolumeSpecName: "utilities") pod "9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" (UID: "9395c2b0-67a7-4a8d-a89a-026dc50f4ae2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.466256 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-kube-api-access-5xwjz" (OuterVolumeSpecName: "kube-api-access-5xwjz") pod "9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" (UID: "9395c2b0-67a7-4a8d-a89a-026dc50f4ae2"). InnerVolumeSpecName "kube-api-access-5xwjz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.531351 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" (UID: "9395c2b0-67a7-4a8d-a89a-026dc50f4ae2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.540134 4787 generic.go:334] "Generic (PLEG): container finished" podID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" containerID="9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900" exitCode=0 Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.540189 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd2pw" event={"ID":"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2","Type":"ContainerDied","Data":"9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900"} Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.540221 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-nd2pw" event={"ID":"9395c2b0-67a7-4a8d-a89a-026dc50f4ae2","Type":"ContainerDied","Data":"18f6ada27532f44902d48d9cd1a138e4e189b7364536b8c37ea88ed69d441d6b"} Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.540242 4787 scope.go:117] "RemoveContainer" containerID="9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.540287 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-nd2pw" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.575425 4787 scope.go:117] "RemoveContainer" containerID="52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.579296 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.579340 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5xwjz\" (UniqueName: \"kubernetes.io/projected/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-kube-api-access-5xwjz\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.579501 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.627357 4787 scope.go:117] "RemoveContainer" containerID="3e9a57d146ec736cde022b223b9bebe29b241303deef9cedf8e6634726812fea" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.633782 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-nd2pw"] Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.645857 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-nd2pw"] Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.678601 4787 scope.go:117] "RemoveContainer" containerID="9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900" Dec 03 19:17:36 crc kubenswrapper[4787]: E1203 19:17:36.679086 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900\": container with ID starting with 9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900 not found: ID does not exist" containerID="9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.679141 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900"} err="failed to get container status \"9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900\": rpc error: code = NotFound desc = could not find container \"9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900\": container with ID starting with 9aec26f3edb3d662cfae65a30013f0d63ecd6768f40ec336f390b05b4ae09900 not found: ID does not exist" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.679176 4787 scope.go:117] "RemoveContainer" containerID="52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708" Dec 03 19:17:36 crc kubenswrapper[4787]: E1203 19:17:36.679600 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708\": container with ID starting with 52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708 not found: ID does not exist" containerID="52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.679632 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708"} err="failed to get container status \"52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708\": rpc error: code = NotFound desc = could not find container \"52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708\": container with ID starting with 52a9a907a1018a8a76fb42d8fbb8c0d84690afa7b02d7c85254c3cfeddf93708 not found: ID does not exist" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.679653 4787 scope.go:117] "RemoveContainer" containerID="3e9a57d146ec736cde022b223b9bebe29b241303deef9cedf8e6634726812fea" Dec 03 19:17:36 crc kubenswrapper[4787]: E1203 19:17:36.681812 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e9a57d146ec736cde022b223b9bebe29b241303deef9cedf8e6634726812fea\": container with ID starting with 3e9a57d146ec736cde022b223b9bebe29b241303deef9cedf8e6634726812fea not found: ID does not exist" containerID="3e9a57d146ec736cde022b223b9bebe29b241303deef9cedf8e6634726812fea" Dec 03 19:17:36 crc kubenswrapper[4787]: I1203 19:17:36.681847 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e9a57d146ec736cde022b223b9bebe29b241303deef9cedf8e6634726812fea"} err="failed to get container status \"3e9a57d146ec736cde022b223b9bebe29b241303deef9cedf8e6634726812fea\": rpc error: code = NotFound desc = could not find container \"3e9a57d146ec736cde022b223b9bebe29b241303deef9cedf8e6634726812fea\": container with ID starting with 3e9a57d146ec736cde022b223b9bebe29b241303deef9cedf8e6634726812fea not found: ID does not exist" Dec 03 19:17:37 crc kubenswrapper[4787]: I1203 19:17:37.789539 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" path="/var/lib/kubelet/pods/9395c2b0-67a7-4a8d-a89a-026dc50f4ae2/volumes" Dec 03 19:17:42 crc kubenswrapper[4787]: I1203 19:17:42.099834 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:42 crc kubenswrapper[4787]: I1203 19:17:42.173526 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nz8q4"] Dec 03 19:17:42 crc kubenswrapper[4787]: I1203 19:17:42.634700 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nz8q4" podUID="d737f2b9-b321-4104-967d-415f3c9c72f1" containerName="registry-server" containerID="cri-o://4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e" gracePeriod=2 Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.241098 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.269064 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bz8hd\" (UniqueName: \"kubernetes.io/projected/d737f2b9-b321-4104-967d-415f3c9c72f1-kube-api-access-bz8hd\") pod \"d737f2b9-b321-4104-967d-415f3c9c72f1\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.269495 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-catalog-content\") pod \"d737f2b9-b321-4104-967d-415f3c9c72f1\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.269654 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-utilities\") pod \"d737f2b9-b321-4104-967d-415f3c9c72f1\" (UID: \"d737f2b9-b321-4104-967d-415f3c9c72f1\") " Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.270363 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-utilities" (OuterVolumeSpecName: "utilities") pod "d737f2b9-b321-4104-967d-415f3c9c72f1" (UID: "d737f2b9-b321-4104-967d-415f3c9c72f1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.270605 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.283578 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d737f2b9-b321-4104-967d-415f3c9c72f1-kube-api-access-bz8hd" (OuterVolumeSpecName: "kube-api-access-bz8hd") pod "d737f2b9-b321-4104-967d-415f3c9c72f1" (UID: "d737f2b9-b321-4104-967d-415f3c9c72f1"). InnerVolumeSpecName "kube-api-access-bz8hd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.324120 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d737f2b9-b321-4104-967d-415f3c9c72f1" (UID: "d737f2b9-b321-4104-967d-415f3c9c72f1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.373889 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d737f2b9-b321-4104-967d-415f3c9c72f1-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.374330 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bz8hd\" (UniqueName: \"kubernetes.io/projected/d737f2b9-b321-4104-967d-415f3c9c72f1-kube-api-access-bz8hd\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.655863 4787 generic.go:334] "Generic (PLEG): container finished" podID="d737f2b9-b321-4104-967d-415f3c9c72f1" containerID="4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e" exitCode=0 Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.656291 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz8q4" event={"ID":"d737f2b9-b321-4104-967d-415f3c9c72f1","Type":"ContainerDied","Data":"4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e"} Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.656514 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nz8q4" event={"ID":"d737f2b9-b321-4104-967d-415f3c9c72f1","Type":"ContainerDied","Data":"cfea4fe240422914b3aa1570594317bc5e399543ba4e9d9085c9314deeb82df9"} Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.656662 4787 scope.go:117] "RemoveContainer" containerID="4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.656975 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nz8q4" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.707751 4787 scope.go:117] "RemoveContainer" containerID="7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.726197 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nz8q4"] Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.741834 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nz8q4"] Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.742929 4787 scope.go:117] "RemoveContainer" containerID="980f85f85cf89934ad5ed8093919a3baaa0afb7df6b23cb4bb83c3a690b889ae" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.780427 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d737f2b9-b321-4104-967d-415f3c9c72f1" path="/var/lib/kubelet/pods/d737f2b9-b321-4104-967d-415f3c9c72f1/volumes" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.813824 4787 scope.go:117] "RemoveContainer" containerID="4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e" Dec 03 19:17:43 crc kubenswrapper[4787]: E1203 19:17:43.814525 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e\": container with ID starting with 4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e not found: ID does not exist" containerID="4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.814560 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e"} err="failed to get container status \"4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e\": rpc error: code = NotFound desc = could not find container \"4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e\": container with ID starting with 4627c26be0ddc5d987a1df0205ad37a530d39303a18ebd95ed0b9706c85a2a4e not found: ID does not exist" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.814578 4787 scope.go:117] "RemoveContainer" containerID="7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a" Dec 03 19:17:43 crc kubenswrapper[4787]: E1203 19:17:43.815093 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a\": container with ID starting with 7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a not found: ID does not exist" containerID="7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.815185 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a"} err="failed to get container status \"7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a\": rpc error: code = NotFound desc = could not find container \"7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a\": container with ID starting with 7b791d6ddeb2010debef36f243425b10dddf594abc469269fc72857f5421370a not found: ID does not exist" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.815249 4787 scope.go:117] "RemoveContainer" containerID="980f85f85cf89934ad5ed8093919a3baaa0afb7df6b23cb4bb83c3a690b889ae" Dec 03 19:17:43 crc kubenswrapper[4787]: E1203 19:17:43.815754 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"980f85f85cf89934ad5ed8093919a3baaa0afb7df6b23cb4bb83c3a690b889ae\": container with ID starting with 980f85f85cf89934ad5ed8093919a3baaa0afb7df6b23cb4bb83c3a690b889ae not found: ID does not exist" containerID="980f85f85cf89934ad5ed8093919a3baaa0afb7df6b23cb4bb83c3a690b889ae" Dec 03 19:17:43 crc kubenswrapper[4787]: I1203 19:17:43.815781 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"980f85f85cf89934ad5ed8093919a3baaa0afb7df6b23cb4bb83c3a690b889ae"} err="failed to get container status \"980f85f85cf89934ad5ed8093919a3baaa0afb7df6b23cb4bb83c3a690b889ae\": rpc error: code = NotFound desc = could not find container \"980f85f85cf89934ad5ed8093919a3baaa0afb7df6b23cb4bb83c3a690b889ae\": container with ID starting with 980f85f85cf89934ad5ed8093919a3baaa0afb7df6b23cb4bb83c3a690b889ae not found: ID does not exist" Dec 03 19:17:44 crc kubenswrapper[4787]: I1203 19:17:44.766954 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:17:44 crc kubenswrapper[4787]: E1203 19:17:44.767584 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:17:50 crc kubenswrapper[4787]: I1203 19:17:50.778173 4787 generic.go:334] "Generic (PLEG): container finished" podID="36d234e6-613c-4745-9e0d-6d4bb3f17162" containerID="85917550bd18b5a593b6b563a5367889774efa3ccba9c571784577eb0b4b2e44" exitCode=0 Dec 03 19:17:50 crc kubenswrapper[4787]: I1203 19:17:50.778310 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/crc-debug-twx2p" event={"ID":"36d234e6-613c-4745-9e0d-6d4bb3f17162","Type":"ContainerDied","Data":"85917550bd18b5a593b6b563a5367889774efa3ccba9c571784577eb0b4b2e44"} Dec 03 19:17:51 crc kubenswrapper[4787]: I1203 19:17:51.920970 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-twx2p" Dec 03 19:17:51 crc kubenswrapper[4787]: I1203 19:17:51.975082 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vj7r9/crc-debug-twx2p"] Dec 03 19:17:51 crc kubenswrapper[4787]: I1203 19:17:51.991407 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vj7r9/crc-debug-twx2p"] Dec 03 19:17:52 crc kubenswrapper[4787]: I1203 19:17:52.012492 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8d8q\" (UniqueName: \"kubernetes.io/projected/36d234e6-613c-4745-9e0d-6d4bb3f17162-kube-api-access-j8d8q\") pod \"36d234e6-613c-4745-9e0d-6d4bb3f17162\" (UID: \"36d234e6-613c-4745-9e0d-6d4bb3f17162\") " Dec 03 19:17:52 crc kubenswrapper[4787]: I1203 19:17:52.012975 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36d234e6-613c-4745-9e0d-6d4bb3f17162-host\") pod \"36d234e6-613c-4745-9e0d-6d4bb3f17162\" (UID: \"36d234e6-613c-4745-9e0d-6d4bb3f17162\") " Dec 03 19:17:52 crc kubenswrapper[4787]: I1203 19:17:52.013075 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/36d234e6-613c-4745-9e0d-6d4bb3f17162-host" (OuterVolumeSpecName: "host") pod "36d234e6-613c-4745-9e0d-6d4bb3f17162" (UID: "36d234e6-613c-4745-9e0d-6d4bb3f17162"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 19:17:52 crc kubenswrapper[4787]: I1203 19:17:52.013966 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/36d234e6-613c-4745-9e0d-6d4bb3f17162-host\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:52 crc kubenswrapper[4787]: I1203 19:17:52.020686 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/36d234e6-613c-4745-9e0d-6d4bb3f17162-kube-api-access-j8d8q" (OuterVolumeSpecName: "kube-api-access-j8d8q") pod "36d234e6-613c-4745-9e0d-6d4bb3f17162" (UID: "36d234e6-613c-4745-9e0d-6d4bb3f17162"). InnerVolumeSpecName "kube-api-access-j8d8q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:17:52 crc kubenswrapper[4787]: I1203 19:17:52.116599 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8d8q\" (UniqueName: \"kubernetes.io/projected/36d234e6-613c-4745-9e0d-6d4bb3f17162-kube-api-access-j8d8q\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:52 crc kubenswrapper[4787]: I1203 19:17:52.802793 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d269538a584bd13fc8caa39f330a363439ad54e50da0f445e9783eccfe33da79" Dec 03 19:17:52 crc kubenswrapper[4787]: I1203 19:17:52.802894 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-twx2p" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.150975 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vj7r9/crc-debug-xxg8x"] Dec 03 19:17:53 crc kubenswrapper[4787]: E1203 19:17:53.151403 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" containerName="registry-server" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.151416 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" containerName="registry-server" Dec 03 19:17:53 crc kubenswrapper[4787]: E1203 19:17:53.151435 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" containerName="extract-content" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.151440 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" containerName="extract-content" Dec 03 19:17:53 crc kubenswrapper[4787]: E1203 19:17:53.151457 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="36d234e6-613c-4745-9e0d-6d4bb3f17162" containerName="container-00" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.151463 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="36d234e6-613c-4745-9e0d-6d4bb3f17162" containerName="container-00" Dec 03 19:17:53 crc kubenswrapper[4787]: E1203 19:17:53.151491 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d737f2b9-b321-4104-967d-415f3c9c72f1" containerName="extract-utilities" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.151499 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d737f2b9-b321-4104-967d-415f3c9c72f1" containerName="extract-utilities" Dec 03 19:17:53 crc kubenswrapper[4787]: E1203 19:17:53.151516 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d737f2b9-b321-4104-967d-415f3c9c72f1" containerName="extract-content" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.151521 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d737f2b9-b321-4104-967d-415f3c9c72f1" containerName="extract-content" Dec 03 19:17:53 crc kubenswrapper[4787]: E1203 19:17:53.151536 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" containerName="extract-utilities" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.151541 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" containerName="extract-utilities" Dec 03 19:17:53 crc kubenswrapper[4787]: E1203 19:17:53.151549 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d737f2b9-b321-4104-967d-415f3c9c72f1" containerName="registry-server" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.151555 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d737f2b9-b321-4104-967d-415f3c9c72f1" containerName="registry-server" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.151747 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d737f2b9-b321-4104-967d-415f3c9c72f1" containerName="registry-server" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.151762 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="36d234e6-613c-4745-9e0d-6d4bb3f17162" containerName="container-00" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.151778 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9395c2b0-67a7-4a8d-a89a-026dc50f4ae2" containerName="registry-server" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.152493 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.244904 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7drfz\" (UniqueName: \"kubernetes.io/projected/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-kube-api-access-7drfz\") pod \"crc-debug-xxg8x\" (UID: \"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762\") " pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.245369 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-host\") pod \"crc-debug-xxg8x\" (UID: \"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762\") " pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.348251 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7drfz\" (UniqueName: \"kubernetes.io/projected/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-kube-api-access-7drfz\") pod \"crc-debug-xxg8x\" (UID: \"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762\") " pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.348564 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-host\") pod \"crc-debug-xxg8x\" (UID: \"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762\") " pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.348725 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-host\") pod \"crc-debug-xxg8x\" (UID: \"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762\") " pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.368233 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7drfz\" (UniqueName: \"kubernetes.io/projected/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-kube-api-access-7drfz\") pod \"crc-debug-xxg8x\" (UID: \"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762\") " pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.482147 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.779050 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="36d234e6-613c-4745-9e0d-6d4bb3f17162" path="/var/lib/kubelet/pods/36d234e6-613c-4745-9e0d-6d4bb3f17162/volumes" Dec 03 19:17:53 crc kubenswrapper[4787]: I1203 19:17:53.813121 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" event={"ID":"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762","Type":"ContainerStarted","Data":"9417afb851c9320c9988891255eca5de8edb010211bddf5829f018dd23befb6c"} Dec 03 19:17:54 crc kubenswrapper[4787]: I1203 19:17:54.830192 4787 generic.go:334] "Generic (PLEG): container finished" podID="d2f53843-8e2f-4bb7-b05d-9a9e2aed9762" containerID="bdadfe39ad0c551d363757c7ea5e1eed12131e7bff312480bfa8a9d0dc6ae0c0" exitCode=0 Dec 03 19:17:54 crc kubenswrapper[4787]: I1203 19:17:54.830279 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" event={"ID":"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762","Type":"ContainerDied","Data":"bdadfe39ad0c551d363757c7ea5e1eed12131e7bff312480bfa8a9d0dc6ae0c0"} Dec 03 19:17:56 crc kubenswrapper[4787]: I1203 19:17:56.013480 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" Dec 03 19:17:56 crc kubenswrapper[4787]: I1203 19:17:56.111678 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7drfz\" (UniqueName: \"kubernetes.io/projected/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-kube-api-access-7drfz\") pod \"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762\" (UID: \"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762\") " Dec 03 19:17:56 crc kubenswrapper[4787]: I1203 19:17:56.111811 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-host\") pod \"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762\" (UID: \"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762\") " Dec 03 19:17:56 crc kubenswrapper[4787]: I1203 19:17:56.111955 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-host" (OuterVolumeSpecName: "host") pod "d2f53843-8e2f-4bb7-b05d-9a9e2aed9762" (UID: "d2f53843-8e2f-4bb7-b05d-9a9e2aed9762"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 19:17:56 crc kubenswrapper[4787]: I1203 19:17:56.112496 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-host\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:56 crc kubenswrapper[4787]: I1203 19:17:56.122299 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-kube-api-access-7drfz" (OuterVolumeSpecName: "kube-api-access-7drfz") pod "d2f53843-8e2f-4bb7-b05d-9a9e2aed9762" (UID: "d2f53843-8e2f-4bb7-b05d-9a9e2aed9762"). InnerVolumeSpecName "kube-api-access-7drfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:17:56 crc kubenswrapper[4787]: I1203 19:17:56.214175 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7drfz\" (UniqueName: \"kubernetes.io/projected/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762-kube-api-access-7drfz\") on node \"crc\" DevicePath \"\"" Dec 03 19:17:56 crc kubenswrapper[4787]: I1203 19:17:56.850824 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" event={"ID":"d2f53843-8e2f-4bb7-b05d-9a9e2aed9762","Type":"ContainerDied","Data":"9417afb851c9320c9988891255eca5de8edb010211bddf5829f018dd23befb6c"} Dec 03 19:17:56 crc kubenswrapper[4787]: I1203 19:17:56.851162 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9417afb851c9320c9988891255eca5de8edb010211bddf5829f018dd23befb6c" Dec 03 19:17:56 crc kubenswrapper[4787]: I1203 19:17:56.850861 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-xxg8x" Dec 03 19:17:57 crc kubenswrapper[4787]: I1203 19:17:57.640488 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vj7r9/crc-debug-xxg8x"] Dec 03 19:17:57 crc kubenswrapper[4787]: I1203 19:17:57.650800 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vj7r9/crc-debug-xxg8x"] Dec 03 19:17:57 crc kubenswrapper[4787]: I1203 19:17:57.765853 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:17:57 crc kubenswrapper[4787]: E1203 19:17:57.766552 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:17:57 crc kubenswrapper[4787]: I1203 19:17:57.780482 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2f53843-8e2f-4bb7-b05d-9a9e2aed9762" path="/var/lib/kubelet/pods/d2f53843-8e2f-4bb7-b05d-9a9e2aed9762/volumes" Dec 03 19:17:58 crc kubenswrapper[4787]: I1203 19:17:58.880608 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vj7r9/crc-debug-85pdt"] Dec 03 19:17:58 crc kubenswrapper[4787]: E1203 19:17:58.881544 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2f53843-8e2f-4bb7-b05d-9a9e2aed9762" containerName="container-00" Dec 03 19:17:58 crc kubenswrapper[4787]: I1203 19:17:58.881562 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2f53843-8e2f-4bb7-b05d-9a9e2aed9762" containerName="container-00" Dec 03 19:17:58 crc kubenswrapper[4787]: I1203 19:17:58.881832 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2f53843-8e2f-4bb7-b05d-9a9e2aed9762" containerName="container-00" Dec 03 19:17:58 crc kubenswrapper[4787]: I1203 19:17:58.882868 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-85pdt" Dec 03 19:17:58 crc kubenswrapper[4787]: I1203 19:17:58.970319 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3caa9d7-390e-452c-ae63-1c6908b70a23-host\") pod \"crc-debug-85pdt\" (UID: \"d3caa9d7-390e-452c-ae63-1c6908b70a23\") " pod="openshift-must-gather-vj7r9/crc-debug-85pdt" Dec 03 19:17:58 crc kubenswrapper[4787]: I1203 19:17:58.970388 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zkd8\" (UniqueName: \"kubernetes.io/projected/d3caa9d7-390e-452c-ae63-1c6908b70a23-kube-api-access-9zkd8\") pod \"crc-debug-85pdt\" (UID: \"d3caa9d7-390e-452c-ae63-1c6908b70a23\") " pod="openshift-must-gather-vj7r9/crc-debug-85pdt" Dec 03 19:17:59 crc kubenswrapper[4787]: I1203 19:17:59.072767 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3caa9d7-390e-452c-ae63-1c6908b70a23-host\") pod \"crc-debug-85pdt\" (UID: \"d3caa9d7-390e-452c-ae63-1c6908b70a23\") " pod="openshift-must-gather-vj7r9/crc-debug-85pdt" Dec 03 19:17:59 crc kubenswrapper[4787]: I1203 19:17:59.072931 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3caa9d7-390e-452c-ae63-1c6908b70a23-host\") pod \"crc-debug-85pdt\" (UID: \"d3caa9d7-390e-452c-ae63-1c6908b70a23\") " pod="openshift-must-gather-vj7r9/crc-debug-85pdt" Dec 03 19:17:59 crc kubenswrapper[4787]: I1203 19:17:59.073312 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zkd8\" (UniqueName: \"kubernetes.io/projected/d3caa9d7-390e-452c-ae63-1c6908b70a23-kube-api-access-9zkd8\") pod \"crc-debug-85pdt\" (UID: \"d3caa9d7-390e-452c-ae63-1c6908b70a23\") " pod="openshift-must-gather-vj7r9/crc-debug-85pdt" Dec 03 19:17:59 crc kubenswrapper[4787]: I1203 19:17:59.095086 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zkd8\" (UniqueName: \"kubernetes.io/projected/d3caa9d7-390e-452c-ae63-1c6908b70a23-kube-api-access-9zkd8\") pod \"crc-debug-85pdt\" (UID: \"d3caa9d7-390e-452c-ae63-1c6908b70a23\") " pod="openshift-must-gather-vj7r9/crc-debug-85pdt" Dec 03 19:17:59 crc kubenswrapper[4787]: I1203 19:17:59.208164 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-85pdt" Dec 03 19:17:59 crc kubenswrapper[4787]: I1203 19:17:59.889731 4787 generic.go:334] "Generic (PLEG): container finished" podID="d3caa9d7-390e-452c-ae63-1c6908b70a23" containerID="839639cdb9abda6a432302982ccdb09bdb9fbfcc6469781fd93db24736177f1a" exitCode=0 Dec 03 19:17:59 crc kubenswrapper[4787]: I1203 19:17:59.890060 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/crc-debug-85pdt" event={"ID":"d3caa9d7-390e-452c-ae63-1c6908b70a23","Type":"ContainerDied","Data":"839639cdb9abda6a432302982ccdb09bdb9fbfcc6469781fd93db24736177f1a"} Dec 03 19:17:59 crc kubenswrapper[4787]: I1203 19:17:59.890113 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/crc-debug-85pdt" event={"ID":"d3caa9d7-390e-452c-ae63-1c6908b70a23","Type":"ContainerStarted","Data":"7af60a8856cb07fa6029fc5ef29421a2f86942f23059280dbf38612c55970ffc"} Dec 03 19:17:59 crc kubenswrapper[4787]: I1203 19:17:59.949780 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vj7r9/crc-debug-85pdt"] Dec 03 19:17:59 crc kubenswrapper[4787]: I1203 19:17:59.963276 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vj7r9/crc-debug-85pdt"] Dec 03 19:18:01 crc kubenswrapper[4787]: I1203 19:18:01.030996 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-85pdt" Dec 03 19:18:01 crc kubenswrapper[4787]: I1203 19:18:01.128456 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zkd8\" (UniqueName: \"kubernetes.io/projected/d3caa9d7-390e-452c-ae63-1c6908b70a23-kube-api-access-9zkd8\") pod \"d3caa9d7-390e-452c-ae63-1c6908b70a23\" (UID: \"d3caa9d7-390e-452c-ae63-1c6908b70a23\") " Dec 03 19:18:01 crc kubenswrapper[4787]: I1203 19:18:01.128506 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3caa9d7-390e-452c-ae63-1c6908b70a23-host\") pod \"d3caa9d7-390e-452c-ae63-1c6908b70a23\" (UID: \"d3caa9d7-390e-452c-ae63-1c6908b70a23\") " Dec 03 19:18:01 crc kubenswrapper[4787]: I1203 19:18:01.128668 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d3caa9d7-390e-452c-ae63-1c6908b70a23-host" (OuterVolumeSpecName: "host") pod "d3caa9d7-390e-452c-ae63-1c6908b70a23" (UID: "d3caa9d7-390e-452c-ae63-1c6908b70a23"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 19:18:01 crc kubenswrapper[4787]: I1203 19:18:01.129319 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d3caa9d7-390e-452c-ae63-1c6908b70a23-host\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:01 crc kubenswrapper[4787]: I1203 19:18:01.142737 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3caa9d7-390e-452c-ae63-1c6908b70a23-kube-api-access-9zkd8" (OuterVolumeSpecName: "kube-api-access-9zkd8") pod "d3caa9d7-390e-452c-ae63-1c6908b70a23" (UID: "d3caa9d7-390e-452c-ae63-1c6908b70a23"). InnerVolumeSpecName "kube-api-access-9zkd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:18:01 crc kubenswrapper[4787]: I1203 19:18:01.231289 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zkd8\" (UniqueName: \"kubernetes.io/projected/d3caa9d7-390e-452c-ae63-1c6908b70a23-kube-api-access-9zkd8\") on node \"crc\" DevicePath \"\"" Dec 03 19:18:01 crc kubenswrapper[4787]: I1203 19:18:01.781079 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3caa9d7-390e-452c-ae63-1c6908b70a23" path="/var/lib/kubelet/pods/d3caa9d7-390e-452c-ae63-1c6908b70a23/volumes" Dec 03 19:18:01 crc kubenswrapper[4787]: I1203 19:18:01.909817 4787 scope.go:117] "RemoveContainer" containerID="839639cdb9abda6a432302982ccdb09bdb9fbfcc6469781fd93db24736177f1a" Dec 03 19:18:01 crc kubenswrapper[4787]: I1203 19:18:01.909862 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/crc-debug-85pdt" Dec 03 19:18:08 crc kubenswrapper[4787]: I1203 19:18:08.766265 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:18:08 crc kubenswrapper[4787]: E1203 19:18:08.766939 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:18:19 crc kubenswrapper[4787]: I1203 19:18:19.767822 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:18:19 crc kubenswrapper[4787]: E1203 19:18:19.769061 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:18:26 crc kubenswrapper[4787]: I1203 19:18:26.238181 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_33d3a3f3-e9c2-42f3-a13c-34be4404eb1a/aodh-api/0.log" Dec 03 19:18:26 crc kubenswrapper[4787]: I1203 19:18:26.523264 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_33d3a3f3-e9c2-42f3-a13c-34be4404eb1a/aodh-evaluator/0.log" Dec 03 19:18:26 crc kubenswrapper[4787]: I1203 19:18:26.528705 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_33d3a3f3-e9c2-42f3-a13c-34be4404eb1a/aodh-notifier/0.log" Dec 03 19:18:26 crc kubenswrapper[4787]: I1203 19:18:26.566853 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_33d3a3f3-e9c2-42f3-a13c-34be4404eb1a/aodh-listener/0.log" Dec 03 19:18:26 crc kubenswrapper[4787]: I1203 19:18:26.739738 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-68677f5b6b-9rwg7_9dc4396d-2f35-42e0-a406-4f0a112fe1cd/barbican-api/0.log" Dec 03 19:18:26 crc kubenswrapper[4787]: I1203 19:18:26.754507 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-68677f5b6b-9rwg7_9dc4396d-2f35-42e0-a406-4f0a112fe1cd/barbican-api-log/0.log" Dec 03 19:18:26 crc kubenswrapper[4787]: I1203 19:18:26.965479 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-78b7c4ccf4-bxfhc_d83ca165-acb3-4e27-b8f3-519897743134/barbican-keystone-listener/0.log" Dec 03 19:18:27 crc kubenswrapper[4787]: I1203 19:18:27.094519 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-78b7c4ccf4-bxfhc_d83ca165-acb3-4e27-b8f3-519897743134/barbican-keystone-listener-log/0.log" Dec 03 19:18:27 crc kubenswrapper[4787]: I1203 19:18:27.354841 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5f6df7d97f-5vmdt_65180fdc-f51b-45e7-ac70-05b9489e1201/barbican-worker/0.log" Dec 03 19:18:27 crc kubenswrapper[4787]: I1203 19:18:27.423602 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx_9490b057-48b6-43c7-ae4f-3d09c1c3e16a/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:27 crc kubenswrapper[4787]: I1203 19:18:27.434555 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5f6df7d97f-5vmdt_65180fdc-f51b-45e7-ac70-05b9489e1201/barbican-worker-log/0.log" Dec 03 19:18:27 crc kubenswrapper[4787]: I1203 19:18:27.656600 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_11460db3-3d3a-426a-9980-f1dd41a84497/proxy-httpd/0.log" Dec 03 19:18:27 crc kubenswrapper[4787]: I1203 19:18:27.679226 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_11460db3-3d3a-426a-9980-f1dd41a84497/ceilometer-central-agent/0.log" Dec 03 19:18:27 crc kubenswrapper[4787]: I1203 19:18:27.703817 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_11460db3-3d3a-426a-9980-f1dd41a84497/ceilometer-notification-agent/0.log" Dec 03 19:18:27 crc kubenswrapper[4787]: I1203 19:18:27.827965 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_11460db3-3d3a-426a-9980-f1dd41a84497/sg-core/0.log" Dec 03 19:18:27 crc kubenswrapper[4787]: I1203 19:18:27.900999 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr_862d8a45-30d2-480c-9aa2-2b8019465b79/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:28 crc kubenswrapper[4787]: I1203 19:18:28.007911 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw_9034ed5d-53f4-4db9-9421-516a5945fce9/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:28 crc kubenswrapper[4787]: I1203 19:18:28.196993 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233/cinder-api-log/0.log" Dec 03 19:18:28 crc kubenswrapper[4787]: I1203 19:18:28.208311 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233/cinder-api/0.log" Dec 03 19:18:28 crc kubenswrapper[4787]: I1203 19:18:28.477459 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_77e12136-eaca-4c6a-9346-da3325061b57/probe/0.log" Dec 03 19:18:28 crc kubenswrapper[4787]: I1203 19:18:28.520742 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9cb81af5-f082-4509-b5a5-b0a49301b75e/cinder-scheduler/0.log" Dec 03 19:18:28 crc kubenswrapper[4787]: I1203 19:18:28.529759 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_77e12136-eaca-4c6a-9346-da3325061b57/cinder-backup/0.log" Dec 03 19:18:28 crc kubenswrapper[4787]: I1203 19:18:28.711619 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9cb81af5-f082-4509-b5a5-b0a49301b75e/probe/0.log" Dec 03 19:18:28 crc kubenswrapper[4787]: I1203 19:18:28.811713 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_0f2fd27e-97a7-4019-98fd-1ed092285098/probe/0.log" Dec 03 19:18:28 crc kubenswrapper[4787]: I1203 19:18:28.813382 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_0f2fd27e-97a7-4019-98fd-1ed092285098/cinder-volume/0.log" Dec 03 19:18:28 crc kubenswrapper[4787]: I1203 19:18:28.963509 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj_6df00b2f-8001-4386-83ed-830a938919e1/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:29 crc kubenswrapper[4787]: I1203 19:18:29.075417 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8_d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:29 crc kubenswrapper[4787]: I1203 19:18:29.176130 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-wsddg_060aca68-9071-46af-a627-ed80ad206d74/init/0.log" Dec 03 19:18:29 crc kubenswrapper[4787]: I1203 19:18:29.470644 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-wsddg_060aca68-9071-46af-a627-ed80ad206d74/init/0.log" Dec 03 19:18:29 crc kubenswrapper[4787]: I1203 19:18:29.503486 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_62231118-050c-48ab-9013-0c07ad92cb6b/glance-httpd/0.log" Dec 03 19:18:29 crc kubenswrapper[4787]: I1203 19:18:29.544495 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-wsddg_060aca68-9071-46af-a627-ed80ad206d74/dnsmasq-dns/0.log" Dec 03 19:18:29 crc kubenswrapper[4787]: I1203 19:18:29.776540 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_62231118-050c-48ab-9013-0c07ad92cb6b/glance-log/0.log" Dec 03 19:18:29 crc kubenswrapper[4787]: I1203 19:18:29.793645 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_21ef4218-4f36-423e-abd5-86b398276be8/glance-httpd/0.log" Dec 03 19:18:29 crc kubenswrapper[4787]: I1203 19:18:29.864509 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_21ef4218-4f36-423e-abd5-86b398276be8/glance-log/0.log" Dec 03 19:18:31 crc kubenswrapper[4787]: I1203 19:18:31.831706 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-57d98bb984-f2bhk_e9019347-02c2-4e52-86a8-07da730bb3b1/heat-engine/0.log" Dec 03 19:18:31 crc kubenswrapper[4787]: I1203 19:18:31.854002 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5d686989bb-9689v_99c71646-510d-4f03-9308-4d0a9ed3c854/horizon/0.log" Dec 03 19:18:32 crc kubenswrapper[4787]: I1203 19:18:32.134405 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7_6d97cc3a-8056-4f1a-98c8-a9dedb9a905d/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:32 crc kubenswrapper[4787]: I1203 19:18:32.475436 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-6z8dd_5f10a9ab-33ee-4938-94f7-870cbdb92c9f/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:32 crc kubenswrapper[4787]: I1203 19:18:32.723390 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5d686989bb-9689v_99c71646-510d-4f03-9308-4d0a9ed3c854/horizon-log/0.log" Dec 03 19:18:33 crc kubenswrapper[4787]: I1203 19:18:33.007522 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29413081-xcmrh_762a8bc2-3aa0-498a-852a-3ede435f23df/keystone-cron/0.log" Dec 03 19:18:33 crc kubenswrapper[4787]: I1203 19:18:33.051655 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-796cccd499-c2cvd_ce3bd00e-d1d3-4086-aa4d-d4a48143f533/heat-cfnapi/0.log" Dec 03 19:18:33 crc kubenswrapper[4787]: I1203 19:18:33.204328 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-74b78cc757-dx9cl_2e92220b-62d7-4d80-8cfa-30a84750f99b/heat-api/0.log" Dec 03 19:18:33 crc kubenswrapper[4787]: I1203 19:18:33.329489 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29413141-r9f29_10d9778c-c2bf-40cf-8934-d584cbc0bad7/keystone-cron/0.log" Dec 03 19:18:33 crc kubenswrapper[4787]: I1203 19:18:33.467227 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_c3f8fcb4-9cbd-4029-81a8-8cada9c7b305/kube-state-metrics/0.log" Dec 03 19:18:33 crc kubenswrapper[4787]: I1203 19:18:33.551451 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-69bdc498c7-wqxdv_c244629d-4ad6-4a6f-9f3d-eaceda01c7e8/keystone-api/0.log" Dec 03 19:18:33 crc kubenswrapper[4787]: I1203 19:18:33.649034 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs_7acc721f-ca1b-4a7c-9e11-4c3b58096a51/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:33 crc kubenswrapper[4787]: I1203 19:18:33.701837 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-p8gvb_4f182e93-6f85-4e36-978f-50a273b371f0/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:33 crc kubenswrapper[4787]: I1203 19:18:33.775897 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:18:33 crc kubenswrapper[4787]: E1203 19:18:33.776264 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:18:33 crc kubenswrapper[4787]: I1203 19:18:33.846675 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_01cfd003-6a74-4b88-a56d-0daa4cf79daf/manila-api-log/0.log" Dec 03 19:18:34 crc kubenswrapper[4787]: I1203 19:18:34.016219 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_01cfd003-6a74-4b88-a56d-0daa4cf79daf/manila-api/0.log" Dec 03 19:18:34 crc kubenswrapper[4787]: I1203 19:18:34.027916 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_5999e180-294a-4f1a-ae8f-e5a7dbf73b0b/probe/0.log" Dec 03 19:18:34 crc kubenswrapper[4787]: I1203 19:18:34.055492 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_5999e180-294a-4f1a-ae8f-e5a7dbf73b0b/manila-scheduler/0.log" Dec 03 19:18:34 crc kubenswrapper[4787]: I1203 19:18:34.228698 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b52259b1-98e9-4791-a981-fc62fdd0e138/probe/0.log" Dec 03 19:18:34 crc kubenswrapper[4787]: I1203 19:18:34.241209 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b52259b1-98e9-4791-a981-fc62fdd0e138/manila-share/0.log" Dec 03 19:18:34 crc kubenswrapper[4787]: I1203 19:18:34.542662 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417/mysqld-exporter/0.log" Dec 03 19:18:34 crc kubenswrapper[4787]: I1203 19:18:34.820002 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr_424ea689-3407-48b5-a7cc-63bced4b0565/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:34 crc kubenswrapper[4787]: I1203 19:18:34.848869 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9fd5cd74f-xdkkz_593e4809-69d6-43bf-ae94-a5d96431ca2f/neutron-api/0.log" Dec 03 19:18:34 crc kubenswrapper[4787]: I1203 19:18:34.858891 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9fd5cd74f-xdkkz_593e4809-69d6-43bf-ae94-a5d96431ca2f/neutron-httpd/0.log" Dec 03 19:18:35 crc kubenswrapper[4787]: I1203 19:18:35.694936 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_42612f1a-6592-480c-b475-865267ee2f9f/nova-cell0-conductor-conductor/0.log" Dec 03 19:18:35 crc kubenswrapper[4787]: I1203 19:18:35.736481 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_70d63cec-994c-4594-8521-c9e8b5252363/nova-cell1-conductor-conductor/0.log" Dec 03 19:18:35 crc kubenswrapper[4787]: I1203 19:18:35.937961 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_98746b52-f9a2-4d4f-ab95-2a1c5d7429ec/nova-api-log/0.log" Dec 03 19:18:36 crc kubenswrapper[4787]: I1203 19:18:36.012496 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_47dc3b7d-f87f-4e03-b969-242a9fdbee2b/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 19:18:36 crc kubenswrapper[4787]: I1203 19:18:36.402664 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk_4fea1a97-fa55-49e0-ab06-88ec531f4792/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:36 crc kubenswrapper[4787]: I1203 19:18:36.489653 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_98746b52-f9a2-4d4f-ab95-2a1c5d7429ec/nova-api-api/0.log" Dec 03 19:18:36 crc kubenswrapper[4787]: I1203 19:18:36.767568 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_be1a822f-d457-417d-9f15-7c3f5b309d7c/nova-metadata-log/0.log" Dec 03 19:18:36 crc kubenswrapper[4787]: I1203 19:18:36.990378 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1a00b714-bec1-411f-a649-c9825253f05e/nova-scheduler-scheduler/0.log" Dec 03 19:18:37 crc kubenswrapper[4787]: I1203 19:18:37.032946 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d207ad18-746f-4012-a399-2876e6d6bc10/mysql-bootstrap/0.log" Dec 03 19:18:37 crc kubenswrapper[4787]: I1203 19:18:37.239728 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d207ad18-746f-4012-a399-2876e6d6bc10/mysql-bootstrap/0.log" Dec 03 19:18:37 crc kubenswrapper[4787]: I1203 19:18:37.318924 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d207ad18-746f-4012-a399-2876e6d6bc10/galera/0.log" Dec 03 19:18:37 crc kubenswrapper[4787]: I1203 19:18:37.516877 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_02a278fe-fdca-423f-a302-59a08cf74ba4/mysql-bootstrap/0.log" Dec 03 19:18:37 crc kubenswrapper[4787]: I1203 19:18:37.719236 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_02a278fe-fdca-423f-a302-59a08cf74ba4/mysql-bootstrap/0.log" Dec 03 19:18:37 crc kubenswrapper[4787]: I1203 19:18:37.771987 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_02a278fe-fdca-423f-a302-59a08cf74ba4/galera/0.log" Dec 03 19:18:38 crc kubenswrapper[4787]: I1203 19:18:38.007663 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ff54f756-3848-48ec-a235-d2814ff8d7f8/openstackclient/0.log" Dec 03 19:18:38 crc kubenswrapper[4787]: I1203 19:18:38.050260 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-mbdkp_cbb0a5a5-509c-4786-8225-22ed73d9ab20/openstack-network-exporter/0.log" Dec 03 19:18:38 crc kubenswrapper[4787]: I1203 19:18:38.328674 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mc6r6_cab2a5b3-1157-45ab-ad4b-828a1638fd78/ovsdb-server-init/0.log" Dec 03 19:18:38 crc kubenswrapper[4787]: I1203 19:18:38.477267 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mc6r6_cab2a5b3-1157-45ab-ad4b-828a1638fd78/ovs-vswitchd/0.log" Dec 03 19:18:38 crc kubenswrapper[4787]: I1203 19:18:38.532010 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mc6r6_cab2a5b3-1157-45ab-ad4b-828a1638fd78/ovsdb-server-init/0.log" Dec 03 19:18:38 crc kubenswrapper[4787]: I1203 19:18:38.584940 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mc6r6_cab2a5b3-1157-45ab-ad4b-828a1638fd78/ovsdb-server/0.log" Dec 03 19:18:38 crc kubenswrapper[4787]: I1203 19:18:38.731651 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-prdc5_2db4ba58-5d6c-42ff-b386-7765170d9113/ovn-controller/0.log" Dec 03 19:18:39 crc kubenswrapper[4787]: I1203 19:18:39.028665 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-cwgpt_4fe91a7c-36b7-4528-8061-f2c47bede56d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:39 crc kubenswrapper[4787]: I1203 19:18:39.169869 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_df89848b-864f-477f-bea0-e34cfcaba0fb/openstack-network-exporter/0.log" Dec 03 19:18:39 crc kubenswrapper[4787]: I1203 19:18:39.262160 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_df89848b-864f-477f-bea0-e34cfcaba0fb/ovn-northd/0.log" Dec 03 19:18:39 crc kubenswrapper[4787]: I1203 19:18:39.396499 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6b71b702-ef57-41ac-8ada-1fe782d5092a/openstack-network-exporter/0.log" Dec 03 19:18:39 crc kubenswrapper[4787]: I1203 19:18:39.448069 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6b71b702-ef57-41ac-8ada-1fe782d5092a/ovsdbserver-nb/0.log" Dec 03 19:18:39 crc kubenswrapper[4787]: I1203 19:18:39.614078 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cd30e7ed-2464-4857-a563-b02f14871d54/openstack-network-exporter/0.log" Dec 03 19:18:39 crc kubenswrapper[4787]: I1203 19:18:39.689763 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cd30e7ed-2464-4857-a563-b02f14871d54/ovsdbserver-sb/0.log" Dec 03 19:18:39 crc kubenswrapper[4787]: I1203 19:18:39.820693 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_be1a822f-d457-417d-9f15-7c3f5b309d7c/nova-metadata-metadata/0.log" Dec 03 19:18:40 crc kubenswrapper[4787]: I1203 19:18:40.071992 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-65d88cc55b-crfnv_ea6b1667-020b-4078-894d-1b944b308802/placement-log/0.log" Dec 03 19:18:40 crc kubenswrapper[4787]: I1203 19:18:40.132259 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7e89b75-951a-44b1-9362-fa52750d7450/init-config-reloader/0.log" Dec 03 19:18:40 crc kubenswrapper[4787]: I1203 19:18:40.174662 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-65d88cc55b-crfnv_ea6b1667-020b-4078-894d-1b944b308802/placement-api/0.log" Dec 03 19:18:40 crc kubenswrapper[4787]: I1203 19:18:40.403788 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7e89b75-951a-44b1-9362-fa52750d7450/init-config-reloader/0.log" Dec 03 19:18:40 crc kubenswrapper[4787]: I1203 19:18:40.414183 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7e89b75-951a-44b1-9362-fa52750d7450/prometheus/0.log" Dec 03 19:18:40 crc kubenswrapper[4787]: I1203 19:18:40.460245 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7e89b75-951a-44b1-9362-fa52750d7450/thanos-sidecar/0.log" Dec 03 19:18:40 crc kubenswrapper[4787]: I1203 19:18:40.494060 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7e89b75-951a-44b1-9362-fa52750d7450/config-reloader/0.log" Dec 03 19:18:40 crc kubenswrapper[4787]: I1203 19:18:40.682377 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9b8eef40-19d4-4640-9517-f603e62e646f/setup-container/0.log" Dec 03 19:18:40 crc kubenswrapper[4787]: I1203 19:18:40.824044 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9b8eef40-19d4-4640-9517-f603e62e646f/setup-container/0.log" Dec 03 19:18:40 crc kubenswrapper[4787]: I1203 19:18:40.842188 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9b8eef40-19d4-4640-9517-f603e62e646f/rabbitmq/0.log" Dec 03 19:18:40 crc kubenswrapper[4787]: I1203 19:18:40.992119 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d774ea58-6127-497a-8948-104e60bfe29b/setup-container/0.log" Dec 03 19:18:41 crc kubenswrapper[4787]: I1203 19:18:41.162335 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d774ea58-6127-497a-8948-104e60bfe29b/setup-container/0.log" Dec 03 19:18:41 crc kubenswrapper[4787]: I1203 19:18:41.181420 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d774ea58-6127-497a-8948-104e60bfe29b/rabbitmq/0.log" Dec 03 19:18:41 crc kubenswrapper[4787]: I1203 19:18:41.205525 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6_ce4c2dba-e3f9-4745-adbe-92d44c7bb025/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:41 crc kubenswrapper[4787]: I1203 19:18:41.459314 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd_be6c98e3-7b90-462f-ac9b-bf1874f97fc8/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:41 crc kubenswrapper[4787]: I1203 19:18:41.581804 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pfw4d_8fbd7464-0bdd-4280-87a4-f4b14b82b057/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:41 crc kubenswrapper[4787]: I1203 19:18:41.742672 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-g87t4_9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac/ssh-known-hosts-edpm-deployment/0.log" Dec 03 19:18:46 crc kubenswrapper[4787]: I1203 19:18:46.927714 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-4fpwj_973215cc-7718-421c-b374-49e64bb9af3e/swift-ring-rebalance/0.log" Dec 03 19:18:46 crc kubenswrapper[4787]: I1203 19:18:46.962568 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65b7898d75-fjwnb_cf9c3742-1e29-4f87-bb53-9a60ab3c14ee/proxy-server/0.log" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.230713 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65b7898d75-fjwnb_cf9c3742-1e29-4f87-bb53-9a60ab3c14ee/proxy-httpd/0.log" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.331509 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/account-auditor/0.log" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.332606 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/account-reaper/0.log" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.547129 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/container-auditor/0.log" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.548438 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/account-replicator/0.log" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.553983 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/account-server/0.log" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.664481 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/container-replicator/0.log" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.767571 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:18:47 crc kubenswrapper[4787]: E1203 19:18:47.767831 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.785614 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/container-updater/0.log" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.789368 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/container-server/0.log" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.871184 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/object-auditor/0.log" Dec 03 19:18:47 crc kubenswrapper[4787]: I1203 19:18:47.896143 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/object-expirer/0.log" Dec 03 19:18:48 crc kubenswrapper[4787]: I1203 19:18:48.031251 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/object-replicator/0.log" Dec 03 19:18:48 crc kubenswrapper[4787]: I1203 19:18:48.046960 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/object-server/0.log" Dec 03 19:18:48 crc kubenswrapper[4787]: I1203 19:18:48.073666 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/rsync/0.log" Dec 03 19:18:48 crc kubenswrapper[4787]: I1203 19:18:48.102754 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/object-updater/0.log" Dec 03 19:18:48 crc kubenswrapper[4787]: I1203 19:18:48.209410 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/swift-recon-cron/0.log" Dec 03 19:18:48 crc kubenswrapper[4787]: I1203 19:18:48.333122 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-c4phf_93b51269-04f3-47d9-b357-9b274a2b3cfe/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:48 crc kubenswrapper[4787]: I1203 19:18:48.475266 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq_69ab81e9-ff7c-40f4-9f76-6f9e06e4caee/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:48 crc kubenswrapper[4787]: I1203 19:18:48.739305 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_dba0eca5-4008-4ce5-b555-d6db725f6466/test-operator-logs-container/0.log" Dec 03 19:18:48 crc kubenswrapper[4787]: I1203 19:18:48.872568 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7_48e931e4-9962-4d61-ad24-62a2a5660510/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:18:49 crc kubenswrapper[4787]: I1203 19:18:49.539555 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_4c367f96-011e-4a7c-89e6-53c9ed2d1c90/tempest-tests-tempest-tests-runner/0.log" Dec 03 19:19:01 crc kubenswrapper[4787]: I1203 19:19:01.766886 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:19:01 crc kubenswrapper[4787]: E1203 19:19:01.767695 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:19:07 crc kubenswrapper[4787]: I1203 19:19:07.788673 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fe87a100-956b-468e-8aa3-164935a19f5f/memcached/0.log" Dec 03 19:19:14 crc kubenswrapper[4787]: I1203 19:19:14.765997 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:19:14 crc kubenswrapper[4787]: E1203 19:19:14.766838 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:19:24 crc kubenswrapper[4787]: I1203 19:19:24.167943 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-bzjxn_43962cfc-342c-49db-83bf-ccde92708a0b/kube-rbac-proxy/0.log" Dec 03 19:19:24 crc kubenswrapper[4787]: I1203 19:19:24.914225 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-mwpr8_02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23/kube-rbac-proxy/0.log" Dec 03 19:19:24 crc kubenswrapper[4787]: I1203 19:19:24.914658 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-fgdm7_6d12a43d-d3da-4b99-b48b-519d660d2527/kube-rbac-proxy/0.log" Dec 03 19:19:24 crc kubenswrapper[4787]: I1203 19:19:24.974744 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-bzjxn_43962cfc-342c-49db-83bf-ccde92708a0b/manager/0.log" Dec 03 19:19:24 crc kubenswrapper[4787]: I1203 19:19:24.987228 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-fgdm7_6d12a43d-d3da-4b99-b48b-519d660d2527/manager/0.log" Dec 03 19:19:25 crc kubenswrapper[4787]: I1203 19:19:25.404257 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-mwpr8_02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23/manager/0.log" Dec 03 19:19:25 crc kubenswrapper[4787]: I1203 19:19:25.503261 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/util/0.log" Dec 03 19:19:25 crc kubenswrapper[4787]: I1203 19:19:25.738226 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/util/0.log" Dec 03 19:19:25 crc kubenswrapper[4787]: I1203 19:19:25.766371 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:19:25 crc kubenswrapper[4787]: E1203 19:19:25.766773 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:19:25 crc kubenswrapper[4787]: I1203 19:19:25.841064 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/pull/0.log" Dec 03 19:19:25 crc kubenswrapper[4787]: I1203 19:19:25.849481 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/pull/0.log" Dec 03 19:19:26 crc kubenswrapper[4787]: I1203 19:19:26.138051 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/util/0.log" Dec 03 19:19:26 crc kubenswrapper[4787]: I1203 19:19:26.142304 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/extract/0.log" Dec 03 19:19:26 crc kubenswrapper[4787]: I1203 19:19:26.181810 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/pull/0.log" Dec 03 19:19:26 crc kubenswrapper[4787]: I1203 19:19:26.414118 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-p425n_355352cc-c3f1-4430-8ff0-55777890e135/kube-rbac-proxy/0.log" Dec 03 19:19:26 crc kubenswrapper[4787]: I1203 19:19:26.476283 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-hr6rk_a6f58c54-821e-4f2b-bce7-1750f64f7ed9/kube-rbac-proxy/0.log" Dec 03 19:19:26 crc kubenswrapper[4787]: I1203 19:19:26.570111 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-p425n_355352cc-c3f1-4430-8ff0-55777890e135/manager/0.log" Dec 03 19:19:26 crc kubenswrapper[4787]: I1203 19:19:26.772549 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-cjbzf_2a789aa2-cc4a-42a9-b1c0-9d1a648fd509/kube-rbac-proxy/0.log" Dec 03 19:19:26 crc kubenswrapper[4787]: I1203 19:19:26.853541 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-cjbzf_2a789aa2-cc4a-42a9-b1c0-9d1a648fd509/manager/0.log" Dec 03 19:19:26 crc kubenswrapper[4787]: I1203 19:19:26.863492 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-hr6rk_a6f58c54-821e-4f2b-bce7-1750f64f7ed9/manager/0.log" Dec 03 19:19:27 crc kubenswrapper[4787]: I1203 19:19:27.095788 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-phcn8_b514a6d2-89f9-451c-af24-c0c5a49cdd8d/kube-rbac-proxy/0.log" Dec 03 19:19:27 crc kubenswrapper[4787]: I1203 19:19:27.235485 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-phcn8_b514a6d2-89f9-451c-af24-c0c5a49cdd8d/manager/0.log" Dec 03 19:19:27 crc kubenswrapper[4787]: I1203 19:19:27.459167 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-bgmsg_8c4a1981-21a3-49ec-a82d-419f8c080fde/kube-rbac-proxy/0.log" Dec 03 19:19:27 crc kubenswrapper[4787]: I1203 19:19:27.469232 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-bgmsg_8c4a1981-21a3-49ec-a82d-419f8c080fde/manager/0.log" Dec 03 19:19:27 crc kubenswrapper[4787]: I1203 19:19:27.622085 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-6txhw_f94890f3-3499-483d-9cda-24a377e982a8/kube-rbac-proxy/0.log" Dec 03 19:19:27 crc kubenswrapper[4787]: I1203 19:19:27.760683 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-6txhw_f94890f3-3499-483d-9cda-24a377e982a8/manager/0.log" Dec 03 19:19:27 crc kubenswrapper[4787]: I1203 19:19:27.878132 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-cd9jr_5c7f6976-b7e2-453d-ad11-6d0e978cd03d/kube-rbac-proxy/0.log" Dec 03 19:19:27 crc kubenswrapper[4787]: I1203 19:19:27.916041 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-cd9jr_5c7f6976-b7e2-453d-ad11-6d0e978cd03d/manager/0.log" Dec 03 19:19:28 crc kubenswrapper[4787]: I1203 19:19:28.056521 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-5m5h7_6384c305-657a-465d-8e48-487fb12cd52f/kube-rbac-proxy/0.log" Dec 03 19:19:28 crc kubenswrapper[4787]: I1203 19:19:28.153840 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-5m5h7_6384c305-657a-465d-8e48-487fb12cd52f/manager/0.log" Dec 03 19:19:28 crc kubenswrapper[4787]: I1203 19:19:28.200737 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-sbcbk_04fbeb64-47f2-4393-8295-0ac7d6db4df3/kube-rbac-proxy/0.log" Dec 03 19:19:28 crc kubenswrapper[4787]: I1203 19:19:28.263244 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-sbcbk_04fbeb64-47f2-4393-8295-0ac7d6db4df3/manager/0.log" Dec 03 19:19:28 crc kubenswrapper[4787]: I1203 19:19:28.435394 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zwmxn_a0326a22-7690-48ed-84f5-01cc9e8331b2/manager/0.log" Dec 03 19:19:28 crc kubenswrapper[4787]: I1203 19:19:28.481050 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zwmxn_a0326a22-7690-48ed-84f5-01cc9e8331b2/kube-rbac-proxy/0.log" Dec 03 19:19:28 crc kubenswrapper[4787]: I1203 19:19:28.601505 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-75r2j_67472bf9-12b8-4463-bc95-dec19e689f36/kube-rbac-proxy/0.log" Dec 03 19:19:28 crc kubenswrapper[4787]: I1203 19:19:28.684789 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-75r2j_67472bf9-12b8-4463-bc95-dec19e689f36/manager/0.log" Dec 03 19:19:28 crc kubenswrapper[4787]: I1203 19:19:28.749787 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz_31c5513f-61f7-4d6c-8136-b212cc171732/kube-rbac-proxy/0.log" Dec 03 19:19:28 crc kubenswrapper[4787]: I1203 19:19:28.787774 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz_31c5513f-61f7-4d6c-8136-b212cc171732/manager/0.log" Dec 03 19:19:29 crc kubenswrapper[4787]: I1203 19:19:29.239150 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5db84b8f89-tm758_3b816649-4a0d-4090-9e9f-df5c5a296777/operator/0.log" Dec 03 19:19:29 crc kubenswrapper[4787]: I1203 19:19:29.255529 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-t5bxl_e5effab2-18ac-4ad1-8d08-618f20509a85/registry-server/0.log" Dec 03 19:19:29 crc kubenswrapper[4787]: I1203 19:19:29.327936 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-xg7kr_226cdcb0-0abc-4a06-9d5c-1d30d9ac8288/kube-rbac-proxy/0.log" Dec 03 19:19:29 crc kubenswrapper[4787]: I1203 19:19:29.607632 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-xg7kr_226cdcb0-0abc-4a06-9d5c-1d30d9ac8288/manager/0.log" Dec 03 19:19:29 crc kubenswrapper[4787]: I1203 19:19:29.609951 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-zgkn8_dd3a3971-adb9-46cb-a0b7-63897e326341/kube-rbac-proxy/0.log" Dec 03 19:19:29 crc kubenswrapper[4787]: I1203 19:19:29.660697 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-zgkn8_dd3a3971-adb9-46cb-a0b7-63897e326341/manager/0.log" Dec 03 19:19:29 crc kubenswrapper[4787]: I1203 19:19:29.814978 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-cnjbp_c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1/operator/0.log" Dec 03 19:19:29 crc kubenswrapper[4787]: I1203 19:19:29.953838 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-vmpfz_c209ad3f-9014-4385-96cb-323dd27b8335/kube-rbac-proxy/0.log" Dec 03 19:19:30 crc kubenswrapper[4787]: I1203 19:19:30.119906 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-66f6f55998-k4kg2_a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173/kube-rbac-proxy/0.log" Dec 03 19:19:30 crc kubenswrapper[4787]: I1203 19:19:30.152529 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-vmpfz_c209ad3f-9014-4385-96cb-323dd27b8335/manager/0.log" Dec 03 19:19:30 crc kubenswrapper[4787]: I1203 19:19:30.373686 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wtvd4_d1364502-c591-47c8-b24c-00c77e3a8d23/kube-rbac-proxy/0.log" Dec 03 19:19:30 crc kubenswrapper[4787]: I1203 19:19:30.383442 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wtvd4_d1364502-c591-47c8-b24c-00c77e3a8d23/manager/0.log" Dec 03 19:19:30 crc kubenswrapper[4787]: I1203 19:19:30.607558 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-2wq6s_8af08022-68dd-4fc4-a052-2cfde5f1ab0a/manager/0.log" Dec 03 19:19:30 crc kubenswrapper[4787]: I1203 19:19:30.630597 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-2wq6s_8af08022-68dd-4fc4-a052-2cfde5f1ab0a/kube-rbac-proxy/0.log" Dec 03 19:19:30 crc kubenswrapper[4787]: I1203 19:19:30.665177 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-66f6f55998-k4kg2_a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173/manager/0.log" Dec 03 19:19:30 crc kubenswrapper[4787]: I1203 19:19:30.736510 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5574b8467d-tg2v7_d9b27877-9ba0-4487-a487-aceafca2075a/manager/0.log" Dec 03 19:19:39 crc kubenswrapper[4787]: I1203 19:19:39.766798 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:19:39 crc kubenswrapper[4787]: E1203 19:19:39.767766 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:19:52 crc kubenswrapper[4787]: I1203 19:19:52.433421 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bj8b6_ee54b630-23ff-4200-aa84-d3aca72f50e8/control-plane-machine-set-operator/0.log" Dec 03 19:19:52 crc kubenswrapper[4787]: I1203 19:19:52.558005 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sm94s_48860c61-124e-43cf-9cf6-fd36f33866f5/kube-rbac-proxy/0.log" Dec 03 19:19:52 crc kubenswrapper[4787]: I1203 19:19:52.638269 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sm94s_48860c61-124e-43cf-9cf6-fd36f33866f5/machine-api-operator/0.log" Dec 03 19:19:53 crc kubenswrapper[4787]: I1203 19:19:53.775833 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:19:53 crc kubenswrapper[4787]: E1203 19:19:53.776383 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:20:07 crc kubenswrapper[4787]: I1203 19:20:07.050418 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-hrd48_249fd549-e9f6-4a50-9b84-b96e46f2791f/cert-manager-controller/0.log" Dec 03 19:20:07 crc kubenswrapper[4787]: I1203 19:20:07.241302 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-f46ld_f5b1552f-1b75-4000-b31e-4d247ea51a65/cert-manager-webhook/0.log" Dec 03 19:20:07 crc kubenswrapper[4787]: I1203 19:20:07.297685 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-r8mcm_c0bb5afc-e67f-42c5-849e-2892c2bf7d3a/cert-manager-cainjector/0.log" Dec 03 19:20:07 crc kubenswrapper[4787]: I1203 19:20:07.766703 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:20:07 crc kubenswrapper[4787]: E1203 19:20:07.766984 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:20:21 crc kubenswrapper[4787]: I1203 19:20:21.766143 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:20:21 crc kubenswrapper[4787]: E1203 19:20:21.766766 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:20:21 crc kubenswrapper[4787]: I1203 19:20:21.871609 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-j2lgj_ce1a6a45-6907-432d-a9c6-3d1f1b9fad52/nmstate-console-plugin/0.log" Dec 03 19:20:22 crc kubenswrapper[4787]: I1203 19:20:22.055283 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-p9brp_2a021e2e-e7b3-44ed-8bc2-3008461b97b3/kube-rbac-proxy/0.log" Dec 03 19:20:22 crc kubenswrapper[4787]: I1203 19:20:22.087387 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-njhbf_f8584654-d601-42cc-98e4-1abd6fdbe848/nmstate-handler/0.log" Dec 03 19:20:22 crc kubenswrapper[4787]: I1203 19:20:22.232732 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-p9brp_2a021e2e-e7b3-44ed-8bc2-3008461b97b3/nmstate-metrics/0.log" Dec 03 19:20:22 crc kubenswrapper[4787]: I1203 19:20:22.357574 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-fhrgj_e4096b9f-51a5-4bb4-860e-0f689a2d9f07/nmstate-operator/0.log" Dec 03 19:20:22 crc kubenswrapper[4787]: I1203 19:20:22.480472 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-b4gp4_2915592e-3cf5-43c3-a142-d89b98274df2/nmstate-webhook/0.log" Dec 03 19:20:36 crc kubenswrapper[4787]: I1203 19:20:36.766637 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:20:36 crc kubenswrapper[4787]: E1203 19:20:36.767251 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:20:37 crc kubenswrapper[4787]: I1203 19:20:37.136525 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5cd867f87d-zwxll_c73f5b31-0cde-406e-800b-153981eeead6/kube-rbac-proxy/0.log" Dec 03 19:20:37 crc kubenswrapper[4787]: I1203 19:20:37.223943 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5cd867f87d-zwxll_c73f5b31-0cde-406e-800b-153981eeead6/manager/0.log" Dec 03 19:20:50 crc kubenswrapper[4787]: I1203 19:20:50.767160 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:20:51 crc kubenswrapper[4787]: E1203 19:20:50.768443 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:20:55 crc kubenswrapper[4787]: I1203 19:20:55.323417 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-z7f2j_28f7de1d-ea99-4b90-bdb9-bc1f04f201ff/cluster-logging-operator/0.log" Dec 03 19:20:55 crc kubenswrapper[4787]: I1203 19:20:55.484151 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-ghzsk_44f70e1d-0f76-4acb-bf6b-1a5ca89469ca/collector/0.log" Dec 03 19:20:55 crc kubenswrapper[4787]: I1203 19:20:55.605100 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_db4243ab-4a52-4991-89a6-96e1ef9a348f/loki-compactor/0.log" Dec 03 19:20:55 crc kubenswrapper[4787]: I1203 19:20:55.717991 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-flghj_e1be9303-41d5-431a-bb49-b0e104ce4625/loki-distributor/0.log" Dec 03 19:20:55 crc kubenswrapper[4787]: I1203 19:20:55.750812 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-57f76f4c5d-dzw77_35486d7a-dcc2-40bf-b2b0-869ecfe46ccc/gateway/0.log" Dec 03 19:20:55 crc kubenswrapper[4787]: I1203 19:20:55.882492 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-57f76f4c5d-dzw77_35486d7a-dcc2-40bf-b2b0-869ecfe46ccc/opa/0.log" Dec 03 19:20:56 crc kubenswrapper[4787]: I1203 19:20:56.116351 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-57f76f4c5d-v45xd_cb532cad-464c-4fe6-99de-b7897de9ac51/gateway/0.log" Dec 03 19:20:56 crc kubenswrapper[4787]: I1203 19:20:56.121466 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-57f76f4c5d-v45xd_cb532cad-464c-4fe6-99de-b7897de9ac51/opa/0.log" Dec 03 19:20:56 crc kubenswrapper[4787]: I1203 19:20:56.294952 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_b4de518a-6a4d-441b-b04c-d35ca9a8b5f9/loki-index-gateway/0.log" Dec 03 19:20:56 crc kubenswrapper[4787]: I1203 19:20:56.475426 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_7da69ee7-38ad-4cce-9bf0-140e7283a775/loki-ingester/0.log" Dec 03 19:20:56 crc kubenswrapper[4787]: I1203 19:20:56.521549 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-xq92s_4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a/loki-querier/0.log" Dec 03 19:20:56 crc kubenswrapper[4787]: I1203 19:20:56.660680 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-zsjlj_b43db9b8-11f8-4fb7-82d2-6b38e34e1a02/loki-query-frontend/0.log" Dec 03 19:21:01 crc kubenswrapper[4787]: I1203 19:21:01.766727 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:21:01 crc kubenswrapper[4787]: E1203 19:21:01.767881 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:21:12 crc kubenswrapper[4787]: I1203 19:21:12.409730 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6rp2q_e1c08eda-0859-42f8-9ca2-f1c4df77038c/kube-rbac-proxy/0.log" Dec 03 19:21:12 crc kubenswrapper[4787]: I1203 19:21:12.655036 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6rp2q_e1c08eda-0859-42f8-9ca2-f1c4df77038c/controller/0.log" Dec 03 19:21:12 crc kubenswrapper[4787]: I1203 19:21:12.798150 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-frr-files/0.log" Dec 03 19:21:12 crc kubenswrapper[4787]: I1203 19:21:12.981419 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-reloader/0.log" Dec 03 19:21:12 crc kubenswrapper[4787]: I1203 19:21:12.998819 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-reloader/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.030470 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-metrics/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.059590 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-frr-files/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.228519 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-frr-files/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.241518 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-reloader/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.250305 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-metrics/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.283522 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-metrics/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.452574 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-reloader/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.482250 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-frr-files/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.512055 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-metrics/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.528160 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/controller/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.660621 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/frr-metrics/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.743494 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/kube-rbac-proxy/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.758722 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/kube-rbac-proxy-frr/0.log" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.773781 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:21:13 crc kubenswrapper[4787]: E1203 19:21:13.774198 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:21:13 crc kubenswrapper[4787]: I1203 19:21:13.868747 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/reloader/0.log" Dec 03 19:21:14 crc kubenswrapper[4787]: I1203 19:21:14.030038 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-kqsb2_55b2904e-d04c-43ef-b054-0ef8636a2316/frr-k8s-webhook-server/0.log" Dec 03 19:21:14 crc kubenswrapper[4787]: I1203 19:21:14.243306 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9b64dcbdc-qzmjb_fe122b60-4167-40b6-8562-1b52112f44f8/manager/0.log" Dec 03 19:21:14 crc kubenswrapper[4787]: I1203 19:21:14.413503 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-86b8bcb77d-fd8v5_200c13aa-bae6-4b29-b6b1-35b625ce2dfc/webhook-server/0.log" Dec 03 19:21:14 crc kubenswrapper[4787]: I1203 19:21:14.503528 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sfcqc_d571a3dc-bd38-4881-857f-e7986b1d90af/kube-rbac-proxy/0.log" Dec 03 19:21:15 crc kubenswrapper[4787]: I1203 19:21:15.306953 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sfcqc_d571a3dc-bd38-4881-857f-e7986b1d90af/speaker/0.log" Dec 03 19:21:15 crc kubenswrapper[4787]: I1203 19:21:15.870977 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/frr/0.log" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.130564 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-jxlgv"] Dec 03 19:21:21 crc kubenswrapper[4787]: E1203 19:21:21.131814 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3caa9d7-390e-452c-ae63-1c6908b70a23" containerName="container-00" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.131829 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3caa9d7-390e-452c-ae63-1c6908b70a23" containerName="container-00" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.132149 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3caa9d7-390e-452c-ae63-1c6908b70a23" containerName="container-00" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.134183 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.160580 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxlgv"] Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.291043 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-utilities\") pod \"redhat-marketplace-jxlgv\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.291283 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-catalog-content\") pod \"redhat-marketplace-jxlgv\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.291611 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qf8r\" (UniqueName: \"kubernetes.io/projected/99216660-54b5-48bf-bb65-1708cd5b2c2b-kube-api-access-5qf8r\") pod \"redhat-marketplace-jxlgv\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.393248 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-utilities\") pod \"redhat-marketplace-jxlgv\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.393552 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-catalog-content\") pod \"redhat-marketplace-jxlgv\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.393703 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qf8r\" (UniqueName: \"kubernetes.io/projected/99216660-54b5-48bf-bb65-1708cd5b2c2b-kube-api-access-5qf8r\") pod \"redhat-marketplace-jxlgv\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.393934 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-utilities\") pod \"redhat-marketplace-jxlgv\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.394308 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-catalog-content\") pod \"redhat-marketplace-jxlgv\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.424946 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qf8r\" (UniqueName: \"kubernetes.io/projected/99216660-54b5-48bf-bb65-1708cd5b2c2b-kube-api-access-5qf8r\") pod \"redhat-marketplace-jxlgv\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.459982 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:21 crc kubenswrapper[4787]: I1203 19:21:21.986060 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxlgv"] Dec 03 19:21:22 crc kubenswrapper[4787]: I1203 19:21:22.289711 4787 generic.go:334] "Generic (PLEG): container finished" podID="99216660-54b5-48bf-bb65-1708cd5b2c2b" containerID="10627d3147ed1bc5c2ba12e7d2b3276926eebeb08e91048dd18d3c44be6d7df8" exitCode=0 Dec 03 19:21:22 crc kubenswrapper[4787]: I1203 19:21:22.289796 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxlgv" event={"ID":"99216660-54b5-48bf-bb65-1708cd5b2c2b","Type":"ContainerDied","Data":"10627d3147ed1bc5c2ba12e7d2b3276926eebeb08e91048dd18d3c44be6d7df8"} Dec 03 19:21:22 crc kubenswrapper[4787]: I1203 19:21:22.289944 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxlgv" event={"ID":"99216660-54b5-48bf-bb65-1708cd5b2c2b","Type":"ContainerStarted","Data":"aca38f2c07e534c312a61bc55187f2c36f4c842e7f6a845d342bf79219ef8037"} Dec 03 19:21:22 crc kubenswrapper[4787]: I1203 19:21:22.292686 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:21:23 crc kubenswrapper[4787]: I1203 19:21:23.303193 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxlgv" event={"ID":"99216660-54b5-48bf-bb65-1708cd5b2c2b","Type":"ContainerStarted","Data":"3182ee864640335de348edb912254c2eeede5cbcb7be6b76cff08b55cfd1a1b6"} Dec 03 19:21:24 crc kubenswrapper[4787]: I1203 19:21:24.314992 4787 generic.go:334] "Generic (PLEG): container finished" podID="99216660-54b5-48bf-bb65-1708cd5b2c2b" containerID="3182ee864640335de348edb912254c2eeede5cbcb7be6b76cff08b55cfd1a1b6" exitCode=0 Dec 03 19:21:24 crc kubenswrapper[4787]: I1203 19:21:24.315167 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxlgv" event={"ID":"99216660-54b5-48bf-bb65-1708cd5b2c2b","Type":"ContainerDied","Data":"3182ee864640335de348edb912254c2eeede5cbcb7be6b76cff08b55cfd1a1b6"} Dec 03 19:21:25 crc kubenswrapper[4787]: I1203 19:21:25.331012 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxlgv" event={"ID":"99216660-54b5-48bf-bb65-1708cd5b2c2b","Type":"ContainerStarted","Data":"60451722a04437acd5d103075149a79838d1fc21c934469333ba9259c8ff1650"} Dec 03 19:21:25 crc kubenswrapper[4787]: I1203 19:21:25.351481 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-jxlgv" podStartSLOduration=1.8925813009999999 podStartE2EDuration="4.351464194s" podCreationTimestamp="2025-12-03 19:21:21 +0000 UTC" firstStartedPulling="2025-12-03 19:21:22.29184188 +0000 UTC m=+7699.109312839" lastFinishedPulling="2025-12-03 19:21:24.750724773 +0000 UTC m=+7701.568195732" observedRunningTime="2025-12-03 19:21:25.348344032 +0000 UTC m=+7702.165815001" watchObservedRunningTime="2025-12-03 19:21:25.351464194 +0000 UTC m=+7702.168935153" Dec 03 19:21:28 crc kubenswrapper[4787]: I1203 19:21:28.768391 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:21:28 crc kubenswrapper[4787]: E1203 19:21:28.769640 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:21:29 crc kubenswrapper[4787]: I1203 19:21:29.340490 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/util/0.log" Dec 03 19:21:29 crc kubenswrapper[4787]: I1203 19:21:29.557545 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/util/0.log" Dec 03 19:21:31 crc kubenswrapper[4787]: I1203 19:21:31.461176 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:31 crc kubenswrapper[4787]: I1203 19:21:31.461583 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:31 crc kubenswrapper[4787]: I1203 19:21:31.519162 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:31 crc kubenswrapper[4787]: I1203 19:21:31.572525 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/pull/0.log" Dec 03 19:21:31 crc kubenswrapper[4787]: I1203 19:21:31.572851 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/pull/0.log" Dec 03 19:21:31 crc kubenswrapper[4787]: I1203 19:21:31.572930 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/pull/0.log" Dec 03 19:21:31 crc kubenswrapper[4787]: I1203 19:21:31.759543 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/util/0.log" Dec 03 19:21:31 crc kubenswrapper[4787]: I1203 19:21:31.811635 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/util/0.log" Dec 03 19:21:31 crc kubenswrapper[4787]: I1203 19:21:31.892043 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/extract/0.log" Dec 03 19:21:32 crc kubenswrapper[4787]: I1203 19:21:32.289648 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/util/0.log" Dec 03 19:21:32 crc kubenswrapper[4787]: I1203 19:21:32.456755 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:32 crc kubenswrapper[4787]: I1203 19:21:32.513050 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxlgv"] Dec 03 19:21:33 crc kubenswrapper[4787]: I1203 19:21:33.499849 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/pull/0.log" Dec 03 19:21:33 crc kubenswrapper[4787]: I1203 19:21:33.499872 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/pull/0.log" Dec 03 19:21:33 crc kubenswrapper[4787]: I1203 19:21:33.706696 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/extract/0.log" Dec 03 19:21:33 crc kubenswrapper[4787]: I1203 19:21:33.932766 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/util/0.log" Dec 03 19:21:33 crc kubenswrapper[4787]: I1203 19:21:33.939477 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/pull/0.log" Dec 03 19:21:33 crc kubenswrapper[4787]: I1203 19:21:33.978971 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/util/0.log" Dec 03 19:21:34 crc kubenswrapper[4787]: I1203 19:21:34.220849 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/util/0.log" Dec 03 19:21:34 crc kubenswrapper[4787]: I1203 19:21:34.326487 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/util/0.log" Dec 03 19:21:34 crc kubenswrapper[4787]: I1203 19:21:34.425240 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-jxlgv" podUID="99216660-54b5-48bf-bb65-1708cd5b2c2b" containerName="registry-server" containerID="cri-o://60451722a04437acd5d103075149a79838d1fc21c934469333ba9259c8ff1650" gracePeriod=2 Dec 03 19:21:34 crc kubenswrapper[4787]: I1203 19:21:34.496899 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/pull/0.log" Dec 03 19:21:34 crc kubenswrapper[4787]: I1203 19:21:34.496966 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/pull/0.log" Dec 03 19:21:34 crc kubenswrapper[4787]: I1203 19:21:34.497163 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/pull/0.log" Dec 03 19:21:34 crc kubenswrapper[4787]: I1203 19:21:34.539859 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/extract/0.log" Dec 03 19:21:34 crc kubenswrapper[4787]: I1203 19:21:34.667879 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/util/0.log" Dec 03 19:21:34 crc kubenswrapper[4787]: I1203 19:21:34.865927 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/pull/0.log" Dec 03 19:21:34 crc kubenswrapper[4787]: I1203 19:21:34.905772 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/pull/0.log" Dec 03 19:21:34 crc kubenswrapper[4787]: I1203 19:21:34.950950 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/util/0.log" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.123627 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/util/0.log" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.194146 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/extract/0.log" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.199367 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/pull/0.log" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.202497 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/util/0.log" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.462932 4787 generic.go:334] "Generic (PLEG): container finished" podID="99216660-54b5-48bf-bb65-1708cd5b2c2b" containerID="60451722a04437acd5d103075149a79838d1fc21c934469333ba9259c8ff1650" exitCode=0 Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.463330 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxlgv" event={"ID":"99216660-54b5-48bf-bb65-1708cd5b2c2b","Type":"ContainerDied","Data":"60451722a04437acd5d103075149a79838d1fc21c934469333ba9259c8ff1650"} Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.477176 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/pull/0.log" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.519398 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/pull/0.log" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.559327 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/util/0.log" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.716399 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.758060 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qf8r\" (UniqueName: \"kubernetes.io/projected/99216660-54b5-48bf-bb65-1708cd5b2c2b-kube-api-access-5qf8r\") pod \"99216660-54b5-48bf-bb65-1708cd5b2c2b\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.758207 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-utilities\") pod \"99216660-54b5-48bf-bb65-1708cd5b2c2b\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.758374 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-catalog-content\") pod \"99216660-54b5-48bf-bb65-1708cd5b2c2b\" (UID: \"99216660-54b5-48bf-bb65-1708cd5b2c2b\") " Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.760086 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-utilities" (OuterVolumeSpecName: "utilities") pod "99216660-54b5-48bf-bb65-1708cd5b2c2b" (UID: "99216660-54b5-48bf-bb65-1708cd5b2c2b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.768998 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/99216660-54b5-48bf-bb65-1708cd5b2c2b-kube-api-access-5qf8r" (OuterVolumeSpecName: "kube-api-access-5qf8r") pod "99216660-54b5-48bf-bb65-1708cd5b2c2b" (UID: "99216660-54b5-48bf-bb65-1708cd5b2c2b"). InnerVolumeSpecName "kube-api-access-5qf8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.790348 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "99216660-54b5-48bf-bb65-1708cd5b2c2b" (UID: "99216660-54b5-48bf-bb65-1708cd5b2c2b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.861219 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qf8r\" (UniqueName: \"kubernetes.io/projected/99216660-54b5-48bf-bb65-1708cd5b2c2b-kube-api-access-5qf8r\") on node \"crc\" DevicePath \"\"" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.861534 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.861543 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/99216660-54b5-48bf-bb65-1708cd5b2c2b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.868654 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/extract/0.log" Dec 03 19:21:35 crc kubenswrapper[4787]: I1203 19:21:35.967489 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/pull/0.log" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.216882 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/util/0.log" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.231109 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-utilities/0.log" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.387392 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-utilities/0.log" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.441035 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-content/0.log" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.467085 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-content/0.log" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.476831 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-jxlgv" event={"ID":"99216660-54b5-48bf-bb65-1708cd5b2c2b","Type":"ContainerDied","Data":"aca38f2c07e534c312a61bc55187f2c36f4c842e7f6a845d342bf79219ef8037"} Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.476898 4787 scope.go:117] "RemoveContainer" containerID="60451722a04437acd5d103075149a79838d1fc21c934469333ba9259c8ff1650" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.477091 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-jxlgv" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.522941 4787 scope.go:117] "RemoveContainer" containerID="3182ee864640335de348edb912254c2eeede5cbcb7be6b76cff08b55cfd1a1b6" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.524323 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxlgv"] Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.542569 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-jxlgv"] Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.556709 4787 scope.go:117] "RemoveContainer" containerID="10627d3147ed1bc5c2ba12e7d2b3276926eebeb08e91048dd18d3c44be6d7df8" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.718444 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-utilities/0.log" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.778392 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-content/0.log" Dec 03 19:21:36 crc kubenswrapper[4787]: I1203 19:21:36.802364 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-utilities/0.log" Dec 03 19:21:37 crc kubenswrapper[4787]: I1203 19:21:37.134348 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-utilities/0.log" Dec 03 19:21:37 crc kubenswrapper[4787]: I1203 19:21:37.146424 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-content/0.log" Dec 03 19:21:37 crc kubenswrapper[4787]: I1203 19:21:37.208127 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-content/0.log" Dec 03 19:21:37 crc kubenswrapper[4787]: I1203 19:21:37.435713 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-content/0.log" Dec 03 19:21:37 crc kubenswrapper[4787]: I1203 19:21:37.492862 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-utilities/0.log" Dec 03 19:21:37 crc kubenswrapper[4787]: I1203 19:21:37.761559 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-278bm_38b06bc8-6c75-46f8-a53b-d95109183306/marketplace-operator/0.log" Dec 03 19:21:37 crc kubenswrapper[4787]: I1203 19:21:37.783504 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="99216660-54b5-48bf-bb65-1708cd5b2c2b" path="/var/lib/kubelet/pods/99216660-54b5-48bf-bb65-1708cd5b2c2b/volumes" Dec 03 19:21:37 crc kubenswrapper[4787]: I1203 19:21:37.955735 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-utilities/0.log" Dec 03 19:21:38 crc kubenswrapper[4787]: I1203 19:21:38.171213 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/registry-server/0.log" Dec 03 19:21:38 crc kubenswrapper[4787]: I1203 19:21:38.191070 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-utilities/0.log" Dec 03 19:21:38 crc kubenswrapper[4787]: I1203 19:21:38.227878 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-content/0.log" Dec 03 19:21:38 crc kubenswrapper[4787]: I1203 19:21:38.258847 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-content/0.log" Dec 03 19:21:38 crc kubenswrapper[4787]: I1203 19:21:38.531006 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-content/0.log" Dec 03 19:21:38 crc kubenswrapper[4787]: I1203 19:21:38.600732 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-utilities/0.log" Dec 03 19:21:39 crc kubenswrapper[4787]: I1203 19:21:39.121825 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-utilities/0.log" Dec 03 19:21:39 crc kubenswrapper[4787]: I1203 19:21:39.330541 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-utilities/0.log" Dec 03 19:21:39 crc kubenswrapper[4787]: I1203 19:21:39.378848 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-content/0.log" Dec 03 19:21:39 crc kubenswrapper[4787]: I1203 19:21:39.381292 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-content/0.log" Dec 03 19:21:39 crc kubenswrapper[4787]: I1203 19:21:39.522981 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-content/0.log" Dec 03 19:21:39 crc kubenswrapper[4787]: I1203 19:21:39.533061 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-utilities/0.log" Dec 03 19:21:42 crc kubenswrapper[4787]: I1203 19:21:42.766716 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:21:42 crc kubenswrapper[4787]: E1203 19:21:42.767436 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:21:43 crc kubenswrapper[4787]: I1203 19:21:43.205743 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/registry-server/0.log" Dec 03 19:21:43 crc kubenswrapper[4787]: I1203 19:21:43.216159 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/registry-server/0.log" Dec 03 19:21:43 crc kubenswrapper[4787]: I1203 19:21:43.230122 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/registry-server/0.log" Dec 03 19:21:54 crc kubenswrapper[4787]: I1203 19:21:54.276342 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-84bqx_d696b783-ca2e-495e-83fe-d33ab80fdf3f/prometheus-operator/0.log" Dec 03 19:21:54 crc kubenswrapper[4787]: I1203 19:21:54.441116 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_1801fa2e-073e-486f-bd71-6abab103dd9f/prometheus-operator-admission-webhook/0.log" Dec 03 19:21:54 crc kubenswrapper[4787]: I1203 19:21:54.471140 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_ede219fc-7a66-484b-831f-6b242694494c/prometheus-operator-admission-webhook/0.log" Dec 03 19:21:54 crc kubenswrapper[4787]: I1203 19:21:54.613624 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-ppms6_bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87/operator/0.log" Dec 03 19:21:54 crc kubenswrapper[4787]: I1203 19:21:54.687481 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-qwtfr_c72c745a-9a1c-4374-b1a7-97279603edf7/observability-ui-dashboards/0.log" Dec 03 19:21:54 crc kubenswrapper[4787]: I1203 19:21:54.795579 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-6cfhr_3651062d-0769-4da5-bb95-7c5987fc2b7e/perses-operator/0.log" Dec 03 19:21:55 crc kubenswrapper[4787]: I1203 19:21:55.766446 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:21:55 crc kubenswrapper[4787]: E1203 19:21:55.767051 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:22:06 crc kubenswrapper[4787]: I1203 19:22:06.766448 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:22:06 crc kubenswrapper[4787]: E1203 19:22:06.767399 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:22:08 crc kubenswrapper[4787]: I1203 19:22:08.045138 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5cd867f87d-zwxll_c73f5b31-0cde-406e-800b-153981eeead6/kube-rbac-proxy/0.log" Dec 03 19:22:08 crc kubenswrapper[4787]: I1203 19:22:08.107465 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5cd867f87d-zwxll_c73f5b31-0cde-406e-800b-153981eeead6/manager/0.log" Dec 03 19:22:17 crc kubenswrapper[4787]: I1203 19:22:17.767261 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:22:17 crc kubenswrapper[4787]: E1203 19:22:17.768607 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:22:30 crc kubenswrapper[4787]: I1203 19:22:30.766673 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:22:31 crc kubenswrapper[4787]: I1203 19:22:31.069748 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"0caa194e8729f36cd086e4c20a6c988ca06fabe2f3e7dc1e38750bed0b6fb608"} Dec 03 19:23:33 crc kubenswrapper[4787]: I1203 19:23:33.020493 4787 scope.go:117] "RemoveContainer" containerID="85917550bd18b5a593b6b563a5367889774efa3ccba9c571784577eb0b4b2e44" Dec 03 19:24:14 crc kubenswrapper[4787]: I1203 19:24:14.462385 4787 generic.go:334] "Generic (PLEG): container finished" podID="a06ae924-3061-4b7e-82fb-345f03981a85" containerID="4d9573ecfa71b743ca24a7534b5a225e65b6bb72bab6df25e461a107813957f4" exitCode=0 Dec 03 19:24:14 crc kubenswrapper[4787]: I1203 19:24:14.462460 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vj7r9/must-gather-l8k5x" event={"ID":"a06ae924-3061-4b7e-82fb-345f03981a85","Type":"ContainerDied","Data":"4d9573ecfa71b743ca24a7534b5a225e65b6bb72bab6df25e461a107813957f4"} Dec 03 19:24:14 crc kubenswrapper[4787]: I1203 19:24:14.464306 4787 scope.go:117] "RemoveContainer" containerID="4d9573ecfa71b743ca24a7534b5a225e65b6bb72bab6df25e461a107813957f4" Dec 03 19:24:14 crc kubenswrapper[4787]: I1203 19:24:14.761829 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vj7r9_must-gather-l8k5x_a06ae924-3061-4b7e-82fb-345f03981a85/gather/0.log" Dec 03 19:24:23 crc kubenswrapper[4787]: I1203 19:24:23.532917 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vj7r9/must-gather-l8k5x"] Dec 03 19:24:23 crc kubenswrapper[4787]: I1203 19:24:23.533861 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-vj7r9/must-gather-l8k5x" podUID="a06ae924-3061-4b7e-82fb-345f03981a85" containerName="copy" containerID="cri-o://5951d2705c1baed53643425dcb9418057aa1bbe056abf41f4d5a80bcdf11e4ba" gracePeriod=2 Dec 03 19:24:23 crc kubenswrapper[4787]: I1203 19:24:23.545428 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vj7r9/must-gather-l8k5x"] Dec 03 19:24:23 crc kubenswrapper[4787]: I1203 19:24:23.677423 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vj7r9_must-gather-l8k5x_a06ae924-3061-4b7e-82fb-345f03981a85/copy/0.log" Dec 03 19:24:23 crc kubenswrapper[4787]: I1203 19:24:23.678173 4787 generic.go:334] "Generic (PLEG): container finished" podID="a06ae924-3061-4b7e-82fb-345f03981a85" containerID="5951d2705c1baed53643425dcb9418057aa1bbe056abf41f4d5a80bcdf11e4ba" exitCode=143 Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.065668 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vj7r9_must-gather-l8k5x_a06ae924-3061-4b7e-82fb-345f03981a85/copy/0.log" Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.066216 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/must-gather-l8k5x" Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.211893 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a06ae924-3061-4b7e-82fb-345f03981a85-must-gather-output\") pod \"a06ae924-3061-4b7e-82fb-345f03981a85\" (UID: \"a06ae924-3061-4b7e-82fb-345f03981a85\") " Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.212267 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvrtk\" (UniqueName: \"kubernetes.io/projected/a06ae924-3061-4b7e-82fb-345f03981a85-kube-api-access-qvrtk\") pod \"a06ae924-3061-4b7e-82fb-345f03981a85\" (UID: \"a06ae924-3061-4b7e-82fb-345f03981a85\") " Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.219763 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a06ae924-3061-4b7e-82fb-345f03981a85-kube-api-access-qvrtk" (OuterVolumeSpecName: "kube-api-access-qvrtk") pod "a06ae924-3061-4b7e-82fb-345f03981a85" (UID: "a06ae924-3061-4b7e-82fb-345f03981a85"). InnerVolumeSpecName "kube-api-access-qvrtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.314779 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvrtk\" (UniqueName: \"kubernetes.io/projected/a06ae924-3061-4b7e-82fb-345f03981a85-kube-api-access-qvrtk\") on node \"crc\" DevicePath \"\"" Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.408079 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a06ae924-3061-4b7e-82fb-345f03981a85-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "a06ae924-3061-4b7e-82fb-345f03981a85" (UID: "a06ae924-3061-4b7e-82fb-345f03981a85"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.416487 4787 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/a06ae924-3061-4b7e-82fb-345f03981a85-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.688164 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vj7r9_must-gather-l8k5x_a06ae924-3061-4b7e-82fb-345f03981a85/copy/0.log" Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.688644 4787 scope.go:117] "RemoveContainer" containerID="5951d2705c1baed53643425dcb9418057aa1bbe056abf41f4d5a80bcdf11e4ba" Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.688690 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vj7r9/must-gather-l8k5x" Dec 03 19:24:24 crc kubenswrapper[4787]: I1203 19:24:24.708765 4787 scope.go:117] "RemoveContainer" containerID="4d9573ecfa71b743ca24a7534b5a225e65b6bb72bab6df25e461a107813957f4" Dec 03 19:24:25 crc kubenswrapper[4787]: I1203 19:24:25.779283 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a06ae924-3061-4b7e-82fb-345f03981a85" path="/var/lib/kubelet/pods/a06ae924-3061-4b7e-82fb-345f03981a85/volumes" Dec 03 19:24:33 crc kubenswrapper[4787]: I1203 19:24:33.118379 4787 scope.go:117] "RemoveContainer" containerID="bdadfe39ad0c551d363757c7ea5e1eed12131e7bff312480bfa8a9d0dc6ae0c0" Dec 03 19:24:48 crc kubenswrapper[4787]: I1203 19:24:48.989401 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:24:48 crc kubenswrapper[4787]: I1203 19:24:48.989981 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:25:18 crc kubenswrapper[4787]: I1203 19:25:18.990105 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:25:18 crc kubenswrapper[4787]: I1203 19:25:18.990671 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:25:48 crc kubenswrapper[4787]: I1203 19:25:48.989425 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:25:48 crc kubenswrapper[4787]: I1203 19:25:48.992952 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:25:48 crc kubenswrapper[4787]: I1203 19:25:48.993168 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 19:25:48 crc kubenswrapper[4787]: I1203 19:25:48.994678 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0caa194e8729f36cd086e4c20a6c988ca06fabe2f3e7dc1e38750bed0b6fb608"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:25:48 crc kubenswrapper[4787]: I1203 19:25:48.994905 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://0caa194e8729f36cd086e4c20a6c988ca06fabe2f3e7dc1e38750bed0b6fb608" gracePeriod=600 Dec 03 19:25:49 crc kubenswrapper[4787]: I1203 19:25:49.775148 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="0caa194e8729f36cd086e4c20a6c988ca06fabe2f3e7dc1e38750bed0b6fb608" exitCode=0 Dec 03 19:25:49 crc kubenswrapper[4787]: I1203 19:25:49.779436 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"0caa194e8729f36cd086e4c20a6c988ca06fabe2f3e7dc1e38750bed0b6fb608"} Dec 03 19:25:49 crc kubenswrapper[4787]: I1203 19:25:49.779504 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294"} Dec 03 19:25:49 crc kubenswrapper[4787]: I1203 19:25:49.779529 4787 scope.go:117] "RemoveContainer" containerID="6881de8cc91b3b1c9ac70ff6727993461a23ee83859cb1ee5c8a588a22128514" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.601288 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-2b98k"] Dec 03 19:27:14 crc kubenswrapper[4787]: E1203 19:27:14.603524 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99216660-54b5-48bf-bb65-1708cd5b2c2b" containerName="extract-content" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.603637 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="99216660-54b5-48bf-bb65-1708cd5b2c2b" containerName="extract-content" Dec 03 19:27:14 crc kubenswrapper[4787]: E1203 19:27:14.603737 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a06ae924-3061-4b7e-82fb-345f03981a85" containerName="copy" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.603816 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a06ae924-3061-4b7e-82fb-345f03981a85" containerName="copy" Dec 03 19:27:14 crc kubenswrapper[4787]: E1203 19:27:14.603957 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a06ae924-3061-4b7e-82fb-345f03981a85" containerName="gather" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.604059 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a06ae924-3061-4b7e-82fb-345f03981a85" containerName="gather" Dec 03 19:27:14 crc kubenswrapper[4787]: E1203 19:27:14.604166 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99216660-54b5-48bf-bb65-1708cd5b2c2b" containerName="extract-utilities" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.604248 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="99216660-54b5-48bf-bb65-1708cd5b2c2b" containerName="extract-utilities" Dec 03 19:27:14 crc kubenswrapper[4787]: E1203 19:27:14.604343 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="99216660-54b5-48bf-bb65-1708cd5b2c2b" containerName="registry-server" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.604421 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="99216660-54b5-48bf-bb65-1708cd5b2c2b" containerName="registry-server" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.604797 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a06ae924-3061-4b7e-82fb-345f03981a85" containerName="gather" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.604906 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="99216660-54b5-48bf-bb65-1708cd5b2c2b" containerName="registry-server" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.604998 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a06ae924-3061-4b7e-82fb-345f03981a85" containerName="copy" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.607102 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.629964 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2b98k"] Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.679842 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-utilities\") pod \"redhat-operators-2b98k\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.679911 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvtks\" (UniqueName: \"kubernetes.io/projected/97774a11-4b55-4b63-a4d0-0add87cbdae7-kube-api-access-qvtks\") pod \"redhat-operators-2b98k\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.679933 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-catalog-content\") pod \"redhat-operators-2b98k\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.782359 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-utilities\") pod \"redhat-operators-2b98k\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.782429 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvtks\" (UniqueName: \"kubernetes.io/projected/97774a11-4b55-4b63-a4d0-0add87cbdae7-kube-api-access-qvtks\") pod \"redhat-operators-2b98k\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.782449 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-catalog-content\") pod \"redhat-operators-2b98k\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.782963 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-catalog-content\") pod \"redhat-operators-2b98k\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.783195 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-utilities\") pod \"redhat-operators-2b98k\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.817823 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvtks\" (UniqueName: \"kubernetes.io/projected/97774a11-4b55-4b63-a4d0-0add87cbdae7-kube-api-access-qvtks\") pod \"redhat-operators-2b98k\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:14 crc kubenswrapper[4787]: I1203 19:27:14.957139 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:15 crc kubenswrapper[4787]: I1203 19:27:15.672822 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-2b98k"] Dec 03 19:27:15 crc kubenswrapper[4787]: I1203 19:27:15.913805 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b98k" event={"ID":"97774a11-4b55-4b63-a4d0-0add87cbdae7","Type":"ContainerStarted","Data":"69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2"} Dec 03 19:27:15 crc kubenswrapper[4787]: I1203 19:27:15.914117 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b98k" event={"ID":"97774a11-4b55-4b63-a4d0-0add87cbdae7","Type":"ContainerStarted","Data":"c2f0d8966249f1c87b45e33d24008926c33a39a527f2a2a26440ed3da5964f2b"} Dec 03 19:27:16 crc kubenswrapper[4787]: I1203 19:27:16.934052 4787 generic.go:334] "Generic (PLEG): container finished" podID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerID="69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2" exitCode=0 Dec 03 19:27:16 crc kubenswrapper[4787]: I1203 19:27:16.934160 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b98k" event={"ID":"97774a11-4b55-4b63-a4d0-0add87cbdae7","Type":"ContainerDied","Data":"69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2"} Dec 03 19:27:16 crc kubenswrapper[4787]: I1203 19:27:16.936825 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:27:17 crc kubenswrapper[4787]: I1203 19:27:17.953121 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b98k" event={"ID":"97774a11-4b55-4b63-a4d0-0add87cbdae7","Type":"ContainerStarted","Data":"ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360"} Dec 03 19:27:23 crc kubenswrapper[4787]: I1203 19:27:23.024388 4787 generic.go:334] "Generic (PLEG): container finished" podID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerID="ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360" exitCode=0 Dec 03 19:27:23 crc kubenswrapper[4787]: I1203 19:27:23.024448 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b98k" event={"ID":"97774a11-4b55-4b63-a4d0-0add87cbdae7","Type":"ContainerDied","Data":"ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360"} Dec 03 19:27:24 crc kubenswrapper[4787]: I1203 19:27:24.041356 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b98k" event={"ID":"97774a11-4b55-4b63-a4d0-0add87cbdae7","Type":"ContainerStarted","Data":"d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4"} Dec 03 19:27:24 crc kubenswrapper[4787]: I1203 19:27:24.060192 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-2b98k" podStartSLOduration=3.506899833 podStartE2EDuration="10.060165828s" podCreationTimestamp="2025-12-03 19:27:14 +0000 UTC" firstStartedPulling="2025-12-03 19:27:16.936618304 +0000 UTC m=+8053.754089263" lastFinishedPulling="2025-12-03 19:27:23.489884289 +0000 UTC m=+8060.307355258" observedRunningTime="2025-12-03 19:27:24.056344138 +0000 UTC m=+8060.873815117" watchObservedRunningTime="2025-12-03 19:27:24.060165828 +0000 UTC m=+8060.877636817" Dec 03 19:27:24 crc kubenswrapper[4787]: I1203 19:27:24.957697 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:24 crc kubenswrapper[4787]: I1203 19:27:24.958115 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:26 crc kubenswrapper[4787]: I1203 19:27:26.009500 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-2b98k" podUID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerName="registry-server" probeResult="failure" output=< Dec 03 19:27:26 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 19:27:26 crc kubenswrapper[4787]: > Dec 03 19:27:32 crc kubenswrapper[4787]: I1203 19:27:32.892450 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-srnnw/must-gather-9xczg"] Dec 03 19:27:32 crc kubenswrapper[4787]: I1203 19:27:32.895410 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/must-gather-9xczg" Dec 03 19:27:32 crc kubenswrapper[4787]: I1203 19:27:32.909934 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-srnnw/must-gather-9xczg"] Dec 03 19:27:32 crc kubenswrapper[4787]: I1203 19:27:32.914715 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-srnnw"/"kube-root-ca.crt" Dec 03 19:27:32 crc kubenswrapper[4787]: I1203 19:27:32.915815 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-srnnw"/"openshift-service-ca.crt" Dec 03 19:27:32 crc kubenswrapper[4787]: I1203 19:27:32.971672 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9b1decf2-f0ad-4149-a10d-874c34248069-must-gather-output\") pod \"must-gather-9xczg\" (UID: \"9b1decf2-f0ad-4149-a10d-874c34248069\") " pod="openshift-must-gather-srnnw/must-gather-9xczg" Dec 03 19:27:32 crc kubenswrapper[4787]: I1203 19:27:32.971830 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rj42q\" (UniqueName: \"kubernetes.io/projected/9b1decf2-f0ad-4149-a10d-874c34248069-kube-api-access-rj42q\") pod \"must-gather-9xczg\" (UID: \"9b1decf2-f0ad-4149-a10d-874c34248069\") " pod="openshift-must-gather-srnnw/must-gather-9xczg" Dec 03 19:27:33 crc kubenswrapper[4787]: I1203 19:27:33.075992 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rj42q\" (UniqueName: \"kubernetes.io/projected/9b1decf2-f0ad-4149-a10d-874c34248069-kube-api-access-rj42q\") pod \"must-gather-9xczg\" (UID: \"9b1decf2-f0ad-4149-a10d-874c34248069\") " pod="openshift-must-gather-srnnw/must-gather-9xczg" Dec 03 19:27:33 crc kubenswrapper[4787]: I1203 19:27:33.076483 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9b1decf2-f0ad-4149-a10d-874c34248069-must-gather-output\") pod \"must-gather-9xczg\" (UID: \"9b1decf2-f0ad-4149-a10d-874c34248069\") " pod="openshift-must-gather-srnnw/must-gather-9xczg" Dec 03 19:27:33 crc kubenswrapper[4787]: I1203 19:27:33.076889 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9b1decf2-f0ad-4149-a10d-874c34248069-must-gather-output\") pod \"must-gather-9xczg\" (UID: \"9b1decf2-f0ad-4149-a10d-874c34248069\") " pod="openshift-must-gather-srnnw/must-gather-9xczg" Dec 03 19:27:33 crc kubenswrapper[4787]: I1203 19:27:33.101056 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rj42q\" (UniqueName: \"kubernetes.io/projected/9b1decf2-f0ad-4149-a10d-874c34248069-kube-api-access-rj42q\") pod \"must-gather-9xczg\" (UID: \"9b1decf2-f0ad-4149-a10d-874c34248069\") " pod="openshift-must-gather-srnnw/must-gather-9xczg" Dec 03 19:27:33 crc kubenswrapper[4787]: I1203 19:27:33.232803 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/must-gather-9xczg" Dec 03 19:27:33 crc kubenswrapper[4787]: I1203 19:27:33.858381 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-srnnw/must-gather-9xczg"] Dec 03 19:27:34 crc kubenswrapper[4787]: I1203 19:27:34.172440 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/must-gather-9xczg" event={"ID":"9b1decf2-f0ad-4149-a10d-874c34248069","Type":"ContainerStarted","Data":"c3abc84527b3f1cf3a11e4fd690d2e06468536c47b52769d3bfc102d9a145807"} Dec 03 19:27:34 crc kubenswrapper[4787]: I1203 19:27:34.172515 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/must-gather-9xczg" event={"ID":"9b1decf2-f0ad-4149-a10d-874c34248069","Type":"ContainerStarted","Data":"8fe8bc4c512f779c255984bdfde6903eefae5556271b5ee9da1f2890afc007d6"} Dec 03 19:27:35 crc kubenswrapper[4787]: I1203 19:27:35.021248 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:35 crc kubenswrapper[4787]: I1203 19:27:35.075504 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:35 crc kubenswrapper[4787]: I1203 19:27:35.190129 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/must-gather-9xczg" event={"ID":"9b1decf2-f0ad-4149-a10d-874c34248069","Type":"ContainerStarted","Data":"36ed78655baaa75f6af15ff9948fea643bce9069835bcf63055438fd251a35b6"} Dec 03 19:27:35 crc kubenswrapper[4787]: I1203 19:27:35.254911 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-srnnw/must-gather-9xczg" podStartSLOduration=3.254895834 podStartE2EDuration="3.254895834s" podCreationTimestamp="2025-12-03 19:27:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 19:27:35.207653151 +0000 UTC m=+8072.025124110" watchObservedRunningTime="2025-12-03 19:27:35.254895834 +0000 UTC m=+8072.072366793" Dec 03 19:27:35 crc kubenswrapper[4787]: I1203 19:27:35.262880 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2b98k"] Dec 03 19:27:36 crc kubenswrapper[4787]: I1203 19:27:36.206044 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-2b98k" podUID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerName="registry-server" containerID="cri-o://d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4" gracePeriod=2 Dec 03 19:27:36 crc kubenswrapper[4787]: I1203 19:27:36.792779 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:36 crc kubenswrapper[4787]: I1203 19:27:36.885120 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-catalog-content\") pod \"97774a11-4b55-4b63-a4d0-0add87cbdae7\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " Dec 03 19:27:36 crc kubenswrapper[4787]: I1203 19:27:36.885220 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvtks\" (UniqueName: \"kubernetes.io/projected/97774a11-4b55-4b63-a4d0-0add87cbdae7-kube-api-access-qvtks\") pod \"97774a11-4b55-4b63-a4d0-0add87cbdae7\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " Dec 03 19:27:36 crc kubenswrapper[4787]: I1203 19:27:36.885499 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-utilities\") pod \"97774a11-4b55-4b63-a4d0-0add87cbdae7\" (UID: \"97774a11-4b55-4b63-a4d0-0add87cbdae7\") " Dec 03 19:27:36 crc kubenswrapper[4787]: I1203 19:27:36.886308 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-utilities" (OuterVolumeSpecName: "utilities") pod "97774a11-4b55-4b63-a4d0-0add87cbdae7" (UID: "97774a11-4b55-4b63-a4d0-0add87cbdae7"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:27:36 crc kubenswrapper[4787]: I1203 19:27:36.887295 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:27:36 crc kubenswrapper[4787]: I1203 19:27:36.916361 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97774a11-4b55-4b63-a4d0-0add87cbdae7-kube-api-access-qvtks" (OuterVolumeSpecName: "kube-api-access-qvtks") pod "97774a11-4b55-4b63-a4d0-0add87cbdae7" (UID: "97774a11-4b55-4b63-a4d0-0add87cbdae7"). InnerVolumeSpecName "kube-api-access-qvtks". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:27:36 crc kubenswrapper[4787]: I1203 19:27:36.989781 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvtks\" (UniqueName: \"kubernetes.io/projected/97774a11-4b55-4b63-a4d0-0add87cbdae7-kube-api-access-qvtks\") on node \"crc\" DevicePath \"\"" Dec 03 19:27:36 crc kubenswrapper[4787]: I1203 19:27:36.994776 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "97774a11-4b55-4b63-a4d0-0add87cbdae7" (UID: "97774a11-4b55-4b63-a4d0-0add87cbdae7"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.091770 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/97774a11-4b55-4b63-a4d0-0add87cbdae7-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.219773 4787 generic.go:334] "Generic (PLEG): container finished" podID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerID="d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4" exitCode=0 Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.219816 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b98k" event={"ID":"97774a11-4b55-4b63-a4d0-0add87cbdae7","Type":"ContainerDied","Data":"d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4"} Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.219844 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-2b98k" event={"ID":"97774a11-4b55-4b63-a4d0-0add87cbdae7","Type":"ContainerDied","Data":"c2f0d8966249f1c87b45e33d24008926c33a39a527f2a2a26440ed3da5964f2b"} Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.219863 4787 scope.go:117] "RemoveContainer" containerID="d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.219862 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-2b98k" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.250457 4787 scope.go:117] "RemoveContainer" containerID="ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.254056 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-2b98k"] Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.276711 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-2b98k"] Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.289247 4787 scope.go:117] "RemoveContainer" containerID="69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.326341 4787 scope.go:117] "RemoveContainer" containerID="d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4" Dec 03 19:27:37 crc kubenswrapper[4787]: E1203 19:27:37.327224 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4\": container with ID starting with d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4 not found: ID does not exist" containerID="d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.327270 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4"} err="failed to get container status \"d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4\": rpc error: code = NotFound desc = could not find container \"d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4\": container with ID starting with d4c2b1e7ddba469df3fe518ce8ea9a405e482c396c3e0b870652467cdbb37dc4 not found: ID does not exist" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.327297 4787 scope.go:117] "RemoveContainer" containerID="ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360" Dec 03 19:27:37 crc kubenswrapper[4787]: E1203 19:27:37.327707 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360\": container with ID starting with ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360 not found: ID does not exist" containerID="ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.327741 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360"} err="failed to get container status \"ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360\": rpc error: code = NotFound desc = could not find container \"ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360\": container with ID starting with ede3a1553863b907090e4500f42fb0bb05bebfdb3431d0756905a595b0bc6360 not found: ID does not exist" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.327761 4787 scope.go:117] "RemoveContainer" containerID="69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2" Dec 03 19:27:37 crc kubenswrapper[4787]: E1203 19:27:37.328110 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2\": container with ID starting with 69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2 not found: ID does not exist" containerID="69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.328160 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2"} err="failed to get container status \"69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2\": rpc error: code = NotFound desc = could not find container \"69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2\": container with ID starting with 69cee4ee43ba19d6e6c5353499bb7dd2d7bb5b017edd249b48dbe48ed4355da2 not found: ID does not exist" Dec 03 19:27:37 crc kubenswrapper[4787]: I1203 19:27:37.779078 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97774a11-4b55-4b63-a4d0-0add87cbdae7" path="/var/lib/kubelet/pods/97774a11-4b55-4b63-a4d0-0add87cbdae7/volumes" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.766103 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-srnnw/crc-debug-v9vzt"] Dec 03 19:27:38 crc kubenswrapper[4787]: E1203 19:27:38.767115 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerName="extract-utilities" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.767131 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerName="extract-utilities" Dec 03 19:27:38 crc kubenswrapper[4787]: E1203 19:27:38.767163 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerName="extract-content" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.767172 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerName="extract-content" Dec 03 19:27:38 crc kubenswrapper[4787]: E1203 19:27:38.767206 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerName="registry-server" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.767214 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerName="registry-server" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.767470 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="97774a11-4b55-4b63-a4d0-0add87cbdae7" containerName="registry-server" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.768409 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-v9vzt" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.783117 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-srnnw"/"default-dockercfg-8qnmx" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.840248 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-host\") pod \"crc-debug-v9vzt\" (UID: \"fb831dbc-aba1-4c2c-a766-ecdbb40e369a\") " pod="openshift-must-gather-srnnw/crc-debug-v9vzt" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.840307 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6ppk\" (UniqueName: \"kubernetes.io/projected/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-kube-api-access-b6ppk\") pod \"crc-debug-v9vzt\" (UID: \"fb831dbc-aba1-4c2c-a766-ecdbb40e369a\") " pod="openshift-must-gather-srnnw/crc-debug-v9vzt" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.942382 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-host\") pod \"crc-debug-v9vzt\" (UID: \"fb831dbc-aba1-4c2c-a766-ecdbb40e369a\") " pod="openshift-must-gather-srnnw/crc-debug-v9vzt" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.942436 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6ppk\" (UniqueName: \"kubernetes.io/projected/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-kube-api-access-b6ppk\") pod \"crc-debug-v9vzt\" (UID: \"fb831dbc-aba1-4c2c-a766-ecdbb40e369a\") " pod="openshift-must-gather-srnnw/crc-debug-v9vzt" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.942514 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-host\") pod \"crc-debug-v9vzt\" (UID: \"fb831dbc-aba1-4c2c-a766-ecdbb40e369a\") " pod="openshift-must-gather-srnnw/crc-debug-v9vzt" Dec 03 19:27:38 crc kubenswrapper[4787]: I1203 19:27:38.979340 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6ppk\" (UniqueName: \"kubernetes.io/projected/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-kube-api-access-b6ppk\") pod \"crc-debug-v9vzt\" (UID: \"fb831dbc-aba1-4c2c-a766-ecdbb40e369a\") " pod="openshift-must-gather-srnnw/crc-debug-v9vzt" Dec 03 19:27:39 crc kubenswrapper[4787]: I1203 19:27:39.084965 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-v9vzt" Dec 03 19:27:39 crc kubenswrapper[4787]: W1203 19:27:39.120567 4787 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfb831dbc_aba1_4c2c_a766_ecdbb40e369a.slice/crio-307edc8850f9e0050ccce7c97bdea83ccef3921ed850459efa7ed9faed7b614f WatchSource:0}: Error finding container 307edc8850f9e0050ccce7c97bdea83ccef3921ed850459efa7ed9faed7b614f: Status 404 returned error can't find the container with id 307edc8850f9e0050ccce7c97bdea83ccef3921ed850459efa7ed9faed7b614f Dec 03 19:27:39 crc kubenswrapper[4787]: I1203 19:27:39.244763 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/crc-debug-v9vzt" event={"ID":"fb831dbc-aba1-4c2c-a766-ecdbb40e369a","Type":"ContainerStarted","Data":"307edc8850f9e0050ccce7c97bdea83ccef3921ed850459efa7ed9faed7b614f"} Dec 03 19:27:40 crc kubenswrapper[4787]: I1203 19:27:40.256888 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/crc-debug-v9vzt" event={"ID":"fb831dbc-aba1-4c2c-a766-ecdbb40e369a","Type":"ContainerStarted","Data":"be2ec87382979b629ed27a341498a17487ecd4c8688e58671f490640106984f3"} Dec 03 19:27:40 crc kubenswrapper[4787]: I1203 19:27:40.278762 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-srnnw/crc-debug-v9vzt" podStartSLOduration=2.278741691 podStartE2EDuration="2.278741691s" podCreationTimestamp="2025-12-03 19:27:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 19:27:40.269184799 +0000 UTC m=+8077.086655778" watchObservedRunningTime="2025-12-03 19:27:40.278741691 +0000 UTC m=+8077.096212650" Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.072811 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-5pdh8"] Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.075543 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.129134 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5pdh8"] Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.174768 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-catalog-content\") pod \"community-operators-5pdh8\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.175334 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-utilities\") pod \"community-operators-5pdh8\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.175534 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mb852\" (UniqueName: \"kubernetes.io/projected/906e6b34-e15e-41ca-a824-c2c419dafa6b-kube-api-access-mb852\") pod \"community-operators-5pdh8\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.277948 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-utilities\") pod \"community-operators-5pdh8\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.278041 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mb852\" (UniqueName: \"kubernetes.io/projected/906e6b34-e15e-41ca-a824-c2c419dafa6b-kube-api-access-mb852\") pod \"community-operators-5pdh8\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.278065 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-catalog-content\") pod \"community-operators-5pdh8\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.278612 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-catalog-content\") pod \"community-operators-5pdh8\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.278843 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-utilities\") pod \"community-operators-5pdh8\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:01 crc kubenswrapper[4787]: I1203 19:28:01.334937 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mb852\" (UniqueName: \"kubernetes.io/projected/906e6b34-e15e-41ca-a824-c2c419dafa6b-kube-api-access-mb852\") pod \"community-operators-5pdh8\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:02 crc kubenswrapper[4787]: I1203 19:28:02.859282 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:03 crc kubenswrapper[4787]: I1203 19:28:03.478405 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-5pdh8"] Dec 03 19:28:03 crc kubenswrapper[4787]: I1203 19:28:03.511201 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdh8" event={"ID":"906e6b34-e15e-41ca-a824-c2c419dafa6b","Type":"ContainerStarted","Data":"d71bb7477fa941f489c710399eba8559888aecf91b0ea706cb3239c4f002fa9b"} Dec 03 19:28:04 crc kubenswrapper[4787]: I1203 19:28:04.521778 4787 generic.go:334] "Generic (PLEG): container finished" podID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerID="5551abf856a4a53317ce6b3edf70f420f49de4e9eb4a2a2ac5303949aef40b5e" exitCode=0 Dec 03 19:28:04 crc kubenswrapper[4787]: I1203 19:28:04.521906 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdh8" event={"ID":"906e6b34-e15e-41ca-a824-c2c419dafa6b","Type":"ContainerDied","Data":"5551abf856a4a53317ce6b3edf70f420f49de4e9eb4a2a2ac5303949aef40b5e"} Dec 03 19:28:06 crc kubenswrapper[4787]: I1203 19:28:06.543515 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdh8" event={"ID":"906e6b34-e15e-41ca-a824-c2c419dafa6b","Type":"ContainerStarted","Data":"a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8"} Dec 03 19:28:07 crc kubenswrapper[4787]: I1203 19:28:07.554081 4787 generic.go:334] "Generic (PLEG): container finished" podID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerID="a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8" exitCode=0 Dec 03 19:28:07 crc kubenswrapper[4787]: I1203 19:28:07.554157 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdh8" event={"ID":"906e6b34-e15e-41ca-a824-c2c419dafa6b","Type":"ContainerDied","Data":"a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8"} Dec 03 19:28:08 crc kubenswrapper[4787]: I1203 19:28:08.568518 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdh8" event={"ID":"906e6b34-e15e-41ca-a824-c2c419dafa6b","Type":"ContainerStarted","Data":"19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0"} Dec 03 19:28:08 crc kubenswrapper[4787]: I1203 19:28:08.592063 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-5pdh8" podStartSLOduration=3.880890029 podStartE2EDuration="7.592041341s" podCreationTimestamp="2025-12-03 19:28:01 +0000 UTC" firstStartedPulling="2025-12-03 19:28:04.524088495 +0000 UTC m=+8101.341559454" lastFinishedPulling="2025-12-03 19:28:08.235239807 +0000 UTC m=+8105.052710766" observedRunningTime="2025-12-03 19:28:08.584033201 +0000 UTC m=+8105.401504170" watchObservedRunningTime="2025-12-03 19:28:08.592041341 +0000 UTC m=+8105.409512300" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.467633 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lx7gh"] Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.472189 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.479836 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lx7gh"] Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.537662 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-utilities\") pod \"certified-operators-lx7gh\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.537715 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dpvzj\" (UniqueName: \"kubernetes.io/projected/c2b0ddd8-dac4-41af-a026-2ec386adafbc-kube-api-access-dpvzj\") pod \"certified-operators-lx7gh\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.537781 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-catalog-content\") pod \"certified-operators-lx7gh\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.649929 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-utilities\") pod \"certified-operators-lx7gh\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.650449 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dpvzj\" (UniqueName: \"kubernetes.io/projected/c2b0ddd8-dac4-41af-a026-2ec386adafbc-kube-api-access-dpvzj\") pod \"certified-operators-lx7gh\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.650540 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-catalog-content\") pod \"certified-operators-lx7gh\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.651819 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-catalog-content\") pod \"certified-operators-lx7gh\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.652147 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-utilities\") pod \"certified-operators-lx7gh\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.680010 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dpvzj\" (UniqueName: \"kubernetes.io/projected/c2b0ddd8-dac4-41af-a026-2ec386adafbc-kube-api-access-dpvzj\") pod \"certified-operators-lx7gh\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.797963 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.861485 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:12 crc kubenswrapper[4787]: I1203 19:28:12.861536 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:13 crc kubenswrapper[4787]: I1203 19:28:13.384280 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lx7gh"] Dec 03 19:28:13 crc kubenswrapper[4787]: I1203 19:28:13.679053 4787 generic.go:334] "Generic (PLEG): container finished" podID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerID="9e7bdbc643c85ea6493d521598aeab6e48164f275d9c144c7b17d714b4c8eedb" exitCode=0 Dec 03 19:28:13 crc kubenswrapper[4787]: I1203 19:28:13.679152 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7gh" event={"ID":"c2b0ddd8-dac4-41af-a026-2ec386adafbc","Type":"ContainerDied","Data":"9e7bdbc643c85ea6493d521598aeab6e48164f275d9c144c7b17d714b4c8eedb"} Dec 03 19:28:13 crc kubenswrapper[4787]: I1203 19:28:13.679368 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7gh" event={"ID":"c2b0ddd8-dac4-41af-a026-2ec386adafbc","Type":"ContainerStarted","Data":"eea9010367318db73edcfa596c6798341bd892a84052666cc2cb95f70cfa4907"} Dec 03 19:28:13 crc kubenswrapper[4787]: I1203 19:28:13.941332 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-5pdh8" podUID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerName="registry-server" probeResult="failure" output=< Dec 03 19:28:13 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 19:28:13 crc kubenswrapper[4787]: > Dec 03 19:28:15 crc kubenswrapper[4787]: I1203 19:28:15.986354 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7gh" event={"ID":"c2b0ddd8-dac4-41af-a026-2ec386adafbc","Type":"ContainerStarted","Data":"bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234"} Dec 03 19:28:18 crc kubenswrapper[4787]: I1203 19:28:18.008055 4787 generic.go:334] "Generic (PLEG): container finished" podID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerID="bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234" exitCode=0 Dec 03 19:28:18 crc kubenswrapper[4787]: I1203 19:28:18.008104 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7gh" event={"ID":"c2b0ddd8-dac4-41af-a026-2ec386adafbc","Type":"ContainerDied","Data":"bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234"} Dec 03 19:28:18 crc kubenswrapper[4787]: I1203 19:28:18.989408 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:28:19 crc kubenswrapper[4787]: I1203 19:28:18.990122 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:28:19 crc kubenswrapper[4787]: I1203 19:28:19.022561 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7gh" event={"ID":"c2b0ddd8-dac4-41af-a026-2ec386adafbc","Type":"ContainerStarted","Data":"aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3"} Dec 03 19:28:19 crc kubenswrapper[4787]: I1203 19:28:19.050604 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lx7gh" podStartSLOduration=2.333768921 podStartE2EDuration="7.050583093s" podCreationTimestamp="2025-12-03 19:28:12 +0000 UTC" firstStartedPulling="2025-12-03 19:28:13.681800062 +0000 UTC m=+8110.499271021" lastFinishedPulling="2025-12-03 19:28:18.398614234 +0000 UTC m=+8115.216085193" observedRunningTime="2025-12-03 19:28:19.040933679 +0000 UTC m=+8115.858404638" watchObservedRunningTime="2025-12-03 19:28:19.050583093 +0000 UTC m=+8115.868054052" Dec 03 19:28:22 crc kubenswrapper[4787]: I1203 19:28:22.798668 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:22 crc kubenswrapper[4787]: I1203 19:28:22.799333 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:22 crc kubenswrapper[4787]: I1203 19:28:22.918487 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:22 crc kubenswrapper[4787]: I1203 19:28:22.977325 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:23 crc kubenswrapper[4787]: I1203 19:28:23.167076 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5pdh8"] Dec 03 19:28:23 crc kubenswrapper[4787]: I1203 19:28:23.848238 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-lx7gh" podUID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerName="registry-server" probeResult="failure" output=< Dec 03 19:28:23 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 19:28:23 crc kubenswrapper[4787]: > Dec 03 19:28:24 crc kubenswrapper[4787]: I1203 19:28:24.074566 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-5pdh8" podUID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerName="registry-server" containerID="cri-o://19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0" gracePeriod=2 Dec 03 19:28:24 crc kubenswrapper[4787]: I1203 19:28:24.634525 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:24 crc kubenswrapper[4787]: I1203 19:28:24.759930 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mb852\" (UniqueName: \"kubernetes.io/projected/906e6b34-e15e-41ca-a824-c2c419dafa6b-kube-api-access-mb852\") pod \"906e6b34-e15e-41ca-a824-c2c419dafa6b\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " Dec 03 19:28:24 crc kubenswrapper[4787]: I1203 19:28:24.760439 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-utilities\") pod \"906e6b34-e15e-41ca-a824-c2c419dafa6b\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " Dec 03 19:28:24 crc kubenswrapper[4787]: I1203 19:28:24.760626 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-catalog-content\") pod \"906e6b34-e15e-41ca-a824-c2c419dafa6b\" (UID: \"906e6b34-e15e-41ca-a824-c2c419dafa6b\") " Dec 03 19:28:24 crc kubenswrapper[4787]: I1203 19:28:24.761718 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-utilities" (OuterVolumeSpecName: "utilities") pod "906e6b34-e15e-41ca-a824-c2c419dafa6b" (UID: "906e6b34-e15e-41ca-a824-c2c419dafa6b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:28:24 crc kubenswrapper[4787]: I1203 19:28:24.780130 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/906e6b34-e15e-41ca-a824-c2c419dafa6b-kube-api-access-mb852" (OuterVolumeSpecName: "kube-api-access-mb852") pod "906e6b34-e15e-41ca-a824-c2c419dafa6b" (UID: "906e6b34-e15e-41ca-a824-c2c419dafa6b"). InnerVolumeSpecName "kube-api-access-mb852". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:28:24 crc kubenswrapper[4787]: I1203 19:28:24.844038 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "906e6b34-e15e-41ca-a824-c2c419dafa6b" (UID: "906e6b34-e15e-41ca-a824-c2c419dafa6b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:28:24 crc kubenswrapper[4787]: I1203 19:28:24.863763 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mb852\" (UniqueName: \"kubernetes.io/projected/906e6b34-e15e-41ca-a824-c2c419dafa6b-kube-api-access-mb852\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:24 crc kubenswrapper[4787]: I1203 19:28:24.863807 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:24 crc kubenswrapper[4787]: I1203 19:28:24.863823 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/906e6b34-e15e-41ca-a824-c2c419dafa6b-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.084738 4787 generic.go:334] "Generic (PLEG): container finished" podID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerID="19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0" exitCode=0 Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.084785 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdh8" event={"ID":"906e6b34-e15e-41ca-a824-c2c419dafa6b","Type":"ContainerDied","Data":"19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0"} Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.084808 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-5pdh8" Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.084827 4787 scope.go:117] "RemoveContainer" containerID="19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0" Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.084815 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-5pdh8" event={"ID":"906e6b34-e15e-41ca-a824-c2c419dafa6b","Type":"ContainerDied","Data":"d71bb7477fa941f489c710399eba8559888aecf91b0ea706cb3239c4f002fa9b"} Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.113324 4787 scope.go:117] "RemoveContainer" containerID="a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8" Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.171169 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-5pdh8"] Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.180940 4787 scope.go:117] "RemoveContainer" containerID="5551abf856a4a53317ce6b3edf70f420f49de4e9eb4a2a2ac5303949aef40b5e" Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.183730 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-5pdh8"] Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.217917 4787 scope.go:117] "RemoveContainer" containerID="19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0" Dec 03 19:28:25 crc kubenswrapper[4787]: E1203 19:28:25.218568 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0\": container with ID starting with 19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0 not found: ID does not exist" containerID="19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0" Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.218617 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0"} err="failed to get container status \"19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0\": rpc error: code = NotFound desc = could not find container \"19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0\": container with ID starting with 19e53908b4f5ee8173a5e059f81ef4e469ee3a106c55e3e46aa5edc2e34e28f0 not found: ID does not exist" Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.218643 4787 scope.go:117] "RemoveContainer" containerID="a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8" Dec 03 19:28:25 crc kubenswrapper[4787]: E1203 19:28:25.219056 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8\": container with ID starting with a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8 not found: ID does not exist" containerID="a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8" Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.219083 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8"} err="failed to get container status \"a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8\": rpc error: code = NotFound desc = could not find container \"a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8\": container with ID starting with a29c52772415981dd8e67f7a1f83b05f84a4da985a03bee7b26059f307f9f6e8 not found: ID does not exist" Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.219096 4787 scope.go:117] "RemoveContainer" containerID="5551abf856a4a53317ce6b3edf70f420f49de4e9eb4a2a2ac5303949aef40b5e" Dec 03 19:28:25 crc kubenswrapper[4787]: E1203 19:28:25.219341 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5551abf856a4a53317ce6b3edf70f420f49de4e9eb4a2a2ac5303949aef40b5e\": container with ID starting with 5551abf856a4a53317ce6b3edf70f420f49de4e9eb4a2a2ac5303949aef40b5e not found: ID does not exist" containerID="5551abf856a4a53317ce6b3edf70f420f49de4e9eb4a2a2ac5303949aef40b5e" Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.219363 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5551abf856a4a53317ce6b3edf70f420f49de4e9eb4a2a2ac5303949aef40b5e"} err="failed to get container status \"5551abf856a4a53317ce6b3edf70f420f49de4e9eb4a2a2ac5303949aef40b5e\": rpc error: code = NotFound desc = could not find container \"5551abf856a4a53317ce6b3edf70f420f49de4e9eb4a2a2ac5303949aef40b5e\": container with ID starting with 5551abf856a4a53317ce6b3edf70f420f49de4e9eb4a2a2ac5303949aef40b5e not found: ID does not exist" Dec 03 19:28:25 crc kubenswrapper[4787]: I1203 19:28:25.779668 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="906e6b34-e15e-41ca-a824-c2c419dafa6b" path="/var/lib/kubelet/pods/906e6b34-e15e-41ca-a824-c2c419dafa6b/volumes" Dec 03 19:28:32 crc kubenswrapper[4787]: I1203 19:28:32.850761 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:32 crc kubenswrapper[4787]: I1203 19:28:32.907850 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:33 crc kubenswrapper[4787]: I1203 19:28:33.101131 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lx7gh"] Dec 03 19:28:34 crc kubenswrapper[4787]: I1203 19:28:34.190944 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lx7gh" podUID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerName="registry-server" containerID="cri-o://aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3" gracePeriod=2 Dec 03 19:28:34 crc kubenswrapper[4787]: I1203 19:28:34.822592 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.015135 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dpvzj\" (UniqueName: \"kubernetes.io/projected/c2b0ddd8-dac4-41af-a026-2ec386adafbc-kube-api-access-dpvzj\") pod \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.016088 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-utilities\") pod \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.016171 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-catalog-content\") pod \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\" (UID: \"c2b0ddd8-dac4-41af-a026-2ec386adafbc\") " Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.017194 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-utilities" (OuterVolumeSpecName: "utilities") pod "c2b0ddd8-dac4-41af-a026-2ec386adafbc" (UID: "c2b0ddd8-dac4-41af-a026-2ec386adafbc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.034103 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c2b0ddd8-dac4-41af-a026-2ec386adafbc-kube-api-access-dpvzj" (OuterVolumeSpecName: "kube-api-access-dpvzj") pod "c2b0ddd8-dac4-41af-a026-2ec386adafbc" (UID: "c2b0ddd8-dac4-41af-a026-2ec386adafbc"). InnerVolumeSpecName "kube-api-access-dpvzj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.071604 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c2b0ddd8-dac4-41af-a026-2ec386adafbc" (UID: "c2b0ddd8-dac4-41af-a026-2ec386adafbc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.119761 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.119810 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c2b0ddd8-dac4-41af-a026-2ec386adafbc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.119826 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dpvzj\" (UniqueName: \"kubernetes.io/projected/c2b0ddd8-dac4-41af-a026-2ec386adafbc-kube-api-access-dpvzj\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.205569 4787 generic.go:334] "Generic (PLEG): container finished" podID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerID="aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3" exitCode=0 Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.205649 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lx7gh" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.205699 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7gh" event={"ID":"c2b0ddd8-dac4-41af-a026-2ec386adafbc","Type":"ContainerDied","Data":"aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3"} Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.207590 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lx7gh" event={"ID":"c2b0ddd8-dac4-41af-a026-2ec386adafbc","Type":"ContainerDied","Data":"eea9010367318db73edcfa596c6798341bd892a84052666cc2cb95f70cfa4907"} Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.207626 4787 scope.go:117] "RemoveContainer" containerID="aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.235921 4787 scope.go:117] "RemoveContainer" containerID="bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.262882 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lx7gh"] Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.276189 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lx7gh"] Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.279079 4787 scope.go:117] "RemoveContainer" containerID="9e7bdbc643c85ea6493d521598aeab6e48164f275d9c144c7b17d714b4c8eedb" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.340408 4787 scope.go:117] "RemoveContainer" containerID="aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3" Dec 03 19:28:35 crc kubenswrapper[4787]: E1203 19:28:35.342036 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3\": container with ID starting with aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3 not found: ID does not exist" containerID="aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.342094 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3"} err="failed to get container status \"aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3\": rpc error: code = NotFound desc = could not find container \"aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3\": container with ID starting with aeb72051e117929a7572c207746381eb7437369a5b487264ca0f4cbcd7e75ad3 not found: ID does not exist" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.342128 4787 scope.go:117] "RemoveContainer" containerID="bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234" Dec 03 19:28:35 crc kubenswrapper[4787]: E1203 19:28:35.342589 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234\": container with ID starting with bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234 not found: ID does not exist" containerID="bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.342641 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234"} err="failed to get container status \"bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234\": rpc error: code = NotFound desc = could not find container \"bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234\": container with ID starting with bd4337e92086c06a59601bcefdd298dbdd15c84b118c5e874ffd6d87f71db234 not found: ID does not exist" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.342673 4787 scope.go:117] "RemoveContainer" containerID="9e7bdbc643c85ea6493d521598aeab6e48164f275d9c144c7b17d714b4c8eedb" Dec 03 19:28:35 crc kubenswrapper[4787]: E1203 19:28:35.343462 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e7bdbc643c85ea6493d521598aeab6e48164f275d9c144c7b17d714b4c8eedb\": container with ID starting with 9e7bdbc643c85ea6493d521598aeab6e48164f275d9c144c7b17d714b4c8eedb not found: ID does not exist" containerID="9e7bdbc643c85ea6493d521598aeab6e48164f275d9c144c7b17d714b4c8eedb" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.343496 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e7bdbc643c85ea6493d521598aeab6e48164f275d9c144c7b17d714b4c8eedb"} err="failed to get container status \"9e7bdbc643c85ea6493d521598aeab6e48164f275d9c144c7b17d714b4c8eedb\": rpc error: code = NotFound desc = could not find container \"9e7bdbc643c85ea6493d521598aeab6e48164f275d9c144c7b17d714b4c8eedb\": container with ID starting with 9e7bdbc643c85ea6493d521598aeab6e48164f275d9c144c7b17d714b4c8eedb not found: ID does not exist" Dec 03 19:28:35 crc kubenswrapper[4787]: I1203 19:28:35.781440 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" path="/var/lib/kubelet/pods/c2b0ddd8-dac4-41af-a026-2ec386adafbc/volumes" Dec 03 19:28:39 crc kubenswrapper[4787]: I1203 19:28:39.255321 4787 generic.go:334] "Generic (PLEG): container finished" podID="fb831dbc-aba1-4c2c-a766-ecdbb40e369a" containerID="be2ec87382979b629ed27a341498a17487ecd4c8688e58671f490640106984f3" exitCode=0 Dec 03 19:28:39 crc kubenswrapper[4787]: I1203 19:28:39.255402 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/crc-debug-v9vzt" event={"ID":"fb831dbc-aba1-4c2c-a766-ecdbb40e369a","Type":"ContainerDied","Data":"be2ec87382979b629ed27a341498a17487ecd4c8688e58671f490640106984f3"} Dec 03 19:28:40 crc kubenswrapper[4787]: I1203 19:28:40.375253 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-v9vzt" Dec 03 19:28:40 crc kubenswrapper[4787]: I1203 19:28:40.411556 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-srnnw/crc-debug-v9vzt"] Dec 03 19:28:40 crc kubenswrapper[4787]: I1203 19:28:40.425062 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-srnnw/crc-debug-v9vzt"] Dec 03 19:28:40 crc kubenswrapper[4787]: I1203 19:28:40.460067 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-host\") pod \"fb831dbc-aba1-4c2c-a766-ecdbb40e369a\" (UID: \"fb831dbc-aba1-4c2c-a766-ecdbb40e369a\") " Dec 03 19:28:40 crc kubenswrapper[4787]: I1203 19:28:40.460192 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-host" (OuterVolumeSpecName: "host") pod "fb831dbc-aba1-4c2c-a766-ecdbb40e369a" (UID: "fb831dbc-aba1-4c2c-a766-ecdbb40e369a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 19:28:40 crc kubenswrapper[4787]: I1203 19:28:40.460403 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b6ppk\" (UniqueName: \"kubernetes.io/projected/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-kube-api-access-b6ppk\") pod \"fb831dbc-aba1-4c2c-a766-ecdbb40e369a\" (UID: \"fb831dbc-aba1-4c2c-a766-ecdbb40e369a\") " Dec 03 19:28:40 crc kubenswrapper[4787]: I1203 19:28:40.461078 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-host\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:40 crc kubenswrapper[4787]: I1203 19:28:40.468194 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-kube-api-access-b6ppk" (OuterVolumeSpecName: "kube-api-access-b6ppk") pod "fb831dbc-aba1-4c2c-a766-ecdbb40e369a" (UID: "fb831dbc-aba1-4c2c-a766-ecdbb40e369a"). InnerVolumeSpecName "kube-api-access-b6ppk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:28:40 crc kubenswrapper[4787]: I1203 19:28:40.563309 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b6ppk\" (UniqueName: \"kubernetes.io/projected/fb831dbc-aba1-4c2c-a766-ecdbb40e369a-kube-api-access-b6ppk\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.285174 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-v9vzt" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.285088 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="307edc8850f9e0050ccce7c97bdea83ccef3921ed850459efa7ed9faed7b614f" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.630460 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-srnnw/crc-debug-fd5l5"] Dec 03 19:28:41 crc kubenswrapper[4787]: E1203 19:28:41.630924 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerName="extract-utilities" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.630936 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerName="extract-utilities" Dec 03 19:28:41 crc kubenswrapper[4787]: E1203 19:28:41.630955 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerName="registry-server" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.630961 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerName="registry-server" Dec 03 19:28:41 crc kubenswrapper[4787]: E1203 19:28:41.630976 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerName="registry-server" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.630983 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerName="registry-server" Dec 03 19:28:41 crc kubenswrapper[4787]: E1203 19:28:41.630995 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerName="extract-utilities" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.631001 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerName="extract-utilities" Dec 03 19:28:41 crc kubenswrapper[4787]: E1203 19:28:41.631036 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerName="extract-content" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.631043 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerName="extract-content" Dec 03 19:28:41 crc kubenswrapper[4787]: E1203 19:28:41.631062 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fb831dbc-aba1-4c2c-a766-ecdbb40e369a" containerName="container-00" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.631067 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="fb831dbc-aba1-4c2c-a766-ecdbb40e369a" containerName="container-00" Dec 03 19:28:41 crc kubenswrapper[4787]: E1203 19:28:41.631081 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerName="extract-content" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.631087 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerName="extract-content" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.631307 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="fb831dbc-aba1-4c2c-a766-ecdbb40e369a" containerName="container-00" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.631580 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="c2b0ddd8-dac4-41af-a026-2ec386adafbc" containerName="registry-server" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.631590 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="906e6b34-e15e-41ca-a824-c2c419dafa6b" containerName="registry-server" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.632592 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-fd5l5" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.634447 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-srnnw"/"default-dockercfg-8qnmx" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.688147 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a218da32-9ae0-4294-aea5-3acdae82c8b8-host\") pod \"crc-debug-fd5l5\" (UID: \"a218da32-9ae0-4294-aea5-3acdae82c8b8\") " pod="openshift-must-gather-srnnw/crc-debug-fd5l5" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.688219 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xklr\" (UniqueName: \"kubernetes.io/projected/a218da32-9ae0-4294-aea5-3acdae82c8b8-kube-api-access-9xklr\") pod \"crc-debug-fd5l5\" (UID: \"a218da32-9ae0-4294-aea5-3acdae82c8b8\") " pod="openshift-must-gather-srnnw/crc-debug-fd5l5" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.779528 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fb831dbc-aba1-4c2c-a766-ecdbb40e369a" path="/var/lib/kubelet/pods/fb831dbc-aba1-4c2c-a766-ecdbb40e369a/volumes" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.791176 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a218da32-9ae0-4294-aea5-3acdae82c8b8-host\") pod \"crc-debug-fd5l5\" (UID: \"a218da32-9ae0-4294-aea5-3acdae82c8b8\") " pod="openshift-must-gather-srnnw/crc-debug-fd5l5" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.791300 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xklr\" (UniqueName: \"kubernetes.io/projected/a218da32-9ae0-4294-aea5-3acdae82c8b8-kube-api-access-9xklr\") pod \"crc-debug-fd5l5\" (UID: \"a218da32-9ae0-4294-aea5-3acdae82c8b8\") " pod="openshift-must-gather-srnnw/crc-debug-fd5l5" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.791346 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a218da32-9ae0-4294-aea5-3acdae82c8b8-host\") pod \"crc-debug-fd5l5\" (UID: \"a218da32-9ae0-4294-aea5-3acdae82c8b8\") " pod="openshift-must-gather-srnnw/crc-debug-fd5l5" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.813275 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xklr\" (UniqueName: \"kubernetes.io/projected/a218da32-9ae0-4294-aea5-3acdae82c8b8-kube-api-access-9xklr\") pod \"crc-debug-fd5l5\" (UID: \"a218da32-9ae0-4294-aea5-3acdae82c8b8\") " pod="openshift-must-gather-srnnw/crc-debug-fd5l5" Dec 03 19:28:41 crc kubenswrapper[4787]: I1203 19:28:41.949602 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-fd5l5" Dec 03 19:28:42 crc kubenswrapper[4787]: I1203 19:28:42.297621 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/crc-debug-fd5l5" event={"ID":"a218da32-9ae0-4294-aea5-3acdae82c8b8","Type":"ContainerStarted","Data":"f9411b2119998f21f1280fc69e0b084d37bf3a215347b9d47cc0baaa83ced7f9"} Dec 03 19:28:42 crc kubenswrapper[4787]: I1203 19:28:42.297946 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/crc-debug-fd5l5" event={"ID":"a218da32-9ae0-4294-aea5-3acdae82c8b8","Type":"ContainerStarted","Data":"db744ec4b96bccb7a4df723dd2bd4d199b638cf97190e69727ac66360af390d8"} Dec 03 19:28:42 crc kubenswrapper[4787]: I1203 19:28:42.317746 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-srnnw/crc-debug-fd5l5" podStartSLOduration=1.317729035 podStartE2EDuration="1.317729035s" podCreationTimestamp="2025-12-03 19:28:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-03 19:28:42.310048833 +0000 UTC m=+8139.127519822" watchObservedRunningTime="2025-12-03 19:28:42.317729035 +0000 UTC m=+8139.135199994" Dec 03 19:28:43 crc kubenswrapper[4787]: I1203 19:28:43.317440 4787 generic.go:334] "Generic (PLEG): container finished" podID="a218da32-9ae0-4294-aea5-3acdae82c8b8" containerID="f9411b2119998f21f1280fc69e0b084d37bf3a215347b9d47cc0baaa83ced7f9" exitCode=0 Dec 03 19:28:43 crc kubenswrapper[4787]: I1203 19:28:43.317715 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/crc-debug-fd5l5" event={"ID":"a218da32-9ae0-4294-aea5-3acdae82c8b8","Type":"ContainerDied","Data":"f9411b2119998f21f1280fc69e0b084d37bf3a215347b9d47cc0baaa83ced7f9"} Dec 03 19:28:44 crc kubenswrapper[4787]: I1203 19:28:44.445151 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-fd5l5" Dec 03 19:28:44 crc kubenswrapper[4787]: I1203 19:28:44.561774 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a218da32-9ae0-4294-aea5-3acdae82c8b8-host\") pod \"a218da32-9ae0-4294-aea5-3acdae82c8b8\" (UID: \"a218da32-9ae0-4294-aea5-3acdae82c8b8\") " Dec 03 19:28:44 crc kubenswrapper[4787]: I1203 19:28:44.562002 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xklr\" (UniqueName: \"kubernetes.io/projected/a218da32-9ae0-4294-aea5-3acdae82c8b8-kube-api-access-9xklr\") pod \"a218da32-9ae0-4294-aea5-3acdae82c8b8\" (UID: \"a218da32-9ae0-4294-aea5-3acdae82c8b8\") " Dec 03 19:28:44 crc kubenswrapper[4787]: I1203 19:28:44.565190 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a218da32-9ae0-4294-aea5-3acdae82c8b8-host" (OuterVolumeSpecName: "host") pod "a218da32-9ae0-4294-aea5-3acdae82c8b8" (UID: "a218da32-9ae0-4294-aea5-3acdae82c8b8"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 19:28:44 crc kubenswrapper[4787]: I1203 19:28:44.567410 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a218da32-9ae0-4294-aea5-3acdae82c8b8-kube-api-access-9xklr" (OuterVolumeSpecName: "kube-api-access-9xklr") pod "a218da32-9ae0-4294-aea5-3acdae82c8b8" (UID: "a218da32-9ae0-4294-aea5-3acdae82c8b8"). InnerVolumeSpecName "kube-api-access-9xklr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:28:44 crc kubenswrapper[4787]: I1203 19:28:44.664892 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xklr\" (UniqueName: \"kubernetes.io/projected/a218da32-9ae0-4294-aea5-3acdae82c8b8-kube-api-access-9xklr\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:44 crc kubenswrapper[4787]: I1203 19:28:44.664932 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/a218da32-9ae0-4294-aea5-3acdae82c8b8-host\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:44 crc kubenswrapper[4787]: I1203 19:28:44.771451 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-srnnw/crc-debug-fd5l5"] Dec 03 19:28:44 crc kubenswrapper[4787]: I1203 19:28:44.799681 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-srnnw/crc-debug-fd5l5"] Dec 03 19:28:45 crc kubenswrapper[4787]: I1203 19:28:45.338053 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="db744ec4b96bccb7a4df723dd2bd4d199b638cf97190e69727ac66360af390d8" Dec 03 19:28:45 crc kubenswrapper[4787]: I1203 19:28:45.338109 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-fd5l5" Dec 03 19:28:45 crc kubenswrapper[4787]: I1203 19:28:45.782699 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a218da32-9ae0-4294-aea5-3acdae82c8b8" path="/var/lib/kubelet/pods/a218da32-9ae0-4294-aea5-3acdae82c8b8/volumes" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.240395 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-srnnw/crc-debug-prdpv"] Dec 03 19:28:46 crc kubenswrapper[4787]: E1203 19:28:46.243567 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a218da32-9ae0-4294-aea5-3acdae82c8b8" containerName="container-00" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.243601 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="a218da32-9ae0-4294-aea5-3acdae82c8b8" containerName="container-00" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.243810 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="a218da32-9ae0-4294-aea5-3acdae82c8b8" containerName="container-00" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.244668 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-prdpv" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.247084 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-srnnw"/"default-dockercfg-8qnmx" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.306608 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e42111fe-d231-4173-b7f0-278ead2d6929-host\") pod \"crc-debug-prdpv\" (UID: \"e42111fe-d231-4173-b7f0-278ead2d6929\") " pod="openshift-must-gather-srnnw/crc-debug-prdpv" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.306804 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfpxz\" (UniqueName: \"kubernetes.io/projected/e42111fe-d231-4173-b7f0-278ead2d6929-kube-api-access-nfpxz\") pod \"crc-debug-prdpv\" (UID: \"e42111fe-d231-4173-b7f0-278ead2d6929\") " pod="openshift-must-gather-srnnw/crc-debug-prdpv" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.408919 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfpxz\" (UniqueName: \"kubernetes.io/projected/e42111fe-d231-4173-b7f0-278ead2d6929-kube-api-access-nfpxz\") pod \"crc-debug-prdpv\" (UID: \"e42111fe-d231-4173-b7f0-278ead2d6929\") " pod="openshift-must-gather-srnnw/crc-debug-prdpv" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.409041 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e42111fe-d231-4173-b7f0-278ead2d6929-host\") pod \"crc-debug-prdpv\" (UID: \"e42111fe-d231-4173-b7f0-278ead2d6929\") " pod="openshift-must-gather-srnnw/crc-debug-prdpv" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.409233 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e42111fe-d231-4173-b7f0-278ead2d6929-host\") pod \"crc-debug-prdpv\" (UID: \"e42111fe-d231-4173-b7f0-278ead2d6929\") " pod="openshift-must-gather-srnnw/crc-debug-prdpv" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.435947 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfpxz\" (UniqueName: \"kubernetes.io/projected/e42111fe-d231-4173-b7f0-278ead2d6929-kube-api-access-nfpxz\") pod \"crc-debug-prdpv\" (UID: \"e42111fe-d231-4173-b7f0-278ead2d6929\") " pod="openshift-must-gather-srnnw/crc-debug-prdpv" Dec 03 19:28:46 crc kubenswrapper[4787]: I1203 19:28:46.566910 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-prdpv" Dec 03 19:28:47 crc kubenswrapper[4787]: I1203 19:28:47.360889 4787 generic.go:334] "Generic (PLEG): container finished" podID="e42111fe-d231-4173-b7f0-278ead2d6929" containerID="2fba621187c579413e557e310c3f3e1eb929b438d570fca2c5e869ab690409e3" exitCode=0 Dec 03 19:28:47 crc kubenswrapper[4787]: I1203 19:28:47.360955 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/crc-debug-prdpv" event={"ID":"e42111fe-d231-4173-b7f0-278ead2d6929","Type":"ContainerDied","Data":"2fba621187c579413e557e310c3f3e1eb929b438d570fca2c5e869ab690409e3"} Dec 03 19:28:47 crc kubenswrapper[4787]: I1203 19:28:47.361291 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/crc-debug-prdpv" event={"ID":"e42111fe-d231-4173-b7f0-278ead2d6929","Type":"ContainerStarted","Data":"7e893736188afbc7f1d1728df607bea02aec4903b26995a9e57897593827a86c"} Dec 03 19:28:47 crc kubenswrapper[4787]: I1203 19:28:47.401542 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-srnnw/crc-debug-prdpv"] Dec 03 19:28:47 crc kubenswrapper[4787]: I1203 19:28:47.411919 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-srnnw/crc-debug-prdpv"] Dec 03 19:28:48 crc kubenswrapper[4787]: I1203 19:28:48.494199 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-prdpv" Dec 03 19:28:48 crc kubenswrapper[4787]: I1203 19:28:48.557566 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e42111fe-d231-4173-b7f0-278ead2d6929-host\") pod \"e42111fe-d231-4173-b7f0-278ead2d6929\" (UID: \"e42111fe-d231-4173-b7f0-278ead2d6929\") " Dec 03 19:28:48 crc kubenswrapper[4787]: I1203 19:28:48.557616 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nfpxz\" (UniqueName: \"kubernetes.io/projected/e42111fe-d231-4173-b7f0-278ead2d6929-kube-api-access-nfpxz\") pod \"e42111fe-d231-4173-b7f0-278ead2d6929\" (UID: \"e42111fe-d231-4173-b7f0-278ead2d6929\") " Dec 03 19:28:48 crc kubenswrapper[4787]: I1203 19:28:48.557678 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e42111fe-d231-4173-b7f0-278ead2d6929-host" (OuterVolumeSpecName: "host") pod "e42111fe-d231-4173-b7f0-278ead2d6929" (UID: "e42111fe-d231-4173-b7f0-278ead2d6929"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 03 19:28:48 crc kubenswrapper[4787]: I1203 19:28:48.558319 4787 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/e42111fe-d231-4173-b7f0-278ead2d6929-host\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:48 crc kubenswrapper[4787]: I1203 19:28:48.563682 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e42111fe-d231-4173-b7f0-278ead2d6929-kube-api-access-nfpxz" (OuterVolumeSpecName: "kube-api-access-nfpxz") pod "e42111fe-d231-4173-b7f0-278ead2d6929" (UID: "e42111fe-d231-4173-b7f0-278ead2d6929"). InnerVolumeSpecName "kube-api-access-nfpxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:28:48 crc kubenswrapper[4787]: I1203 19:28:48.660999 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nfpxz\" (UniqueName: \"kubernetes.io/projected/e42111fe-d231-4173-b7f0-278ead2d6929-kube-api-access-nfpxz\") on node \"crc\" DevicePath \"\"" Dec 03 19:28:48 crc kubenswrapper[4787]: I1203 19:28:48.990268 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:28:48 crc kubenswrapper[4787]: I1203 19:28:48.990327 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:28:49 crc kubenswrapper[4787]: I1203 19:28:49.380272 4787 scope.go:117] "RemoveContainer" containerID="2fba621187c579413e557e310c3f3e1eb929b438d570fca2c5e869ab690409e3" Dec 03 19:28:49 crc kubenswrapper[4787]: I1203 19:28:49.380569 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/crc-debug-prdpv" Dec 03 19:28:49 crc kubenswrapper[4787]: I1203 19:28:49.779928 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e42111fe-d231-4173-b7f0-278ead2d6929" path="/var/lib/kubelet/pods/e42111fe-d231-4173-b7f0-278ead2d6929/volumes" Dec 03 19:29:18 crc kubenswrapper[4787]: I1203 19:29:18.989570 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:29:18 crc kubenswrapper[4787]: I1203 19:29:18.990199 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:29:18 crc kubenswrapper[4787]: I1203 19:29:18.990258 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 19:29:18 crc kubenswrapper[4787]: I1203 19:29:18.991213 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:29:18 crc kubenswrapper[4787]: I1203 19:29:18.991291 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" gracePeriod=600 Dec 03 19:29:19 crc kubenswrapper[4787]: E1203 19:29:19.113893 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:29:19 crc kubenswrapper[4787]: I1203 19:29:19.717159 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" exitCode=0 Dec 03 19:29:19 crc kubenswrapper[4787]: I1203 19:29:19.717201 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294"} Dec 03 19:29:19 crc kubenswrapper[4787]: I1203 19:29:19.717233 4787 scope.go:117] "RemoveContainer" containerID="0caa194e8729f36cd086e4c20a6c988ca06fabe2f3e7dc1e38750bed0b6fb608" Dec 03 19:29:19 crc kubenswrapper[4787]: I1203 19:29:19.718095 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:29:19 crc kubenswrapper[4787]: E1203 19:29:19.718539 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:29:33 crc kubenswrapper[4787]: I1203 19:29:33.773791 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:29:33 crc kubenswrapper[4787]: E1203 19:29:33.774622 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:29:45 crc kubenswrapper[4787]: I1203 19:29:45.018827 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_33d3a3f3-e9c2-42f3-a13c-34be4404eb1a/aodh-api/0.log" Dec 03 19:29:45 crc kubenswrapper[4787]: I1203 19:29:45.203261 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_33d3a3f3-e9c2-42f3-a13c-34be4404eb1a/aodh-listener/0.log" Dec 03 19:29:45 crc kubenswrapper[4787]: I1203 19:29:45.213226 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_33d3a3f3-e9c2-42f3-a13c-34be4404eb1a/aodh-evaluator/0.log" Dec 03 19:29:45 crc kubenswrapper[4787]: I1203 19:29:45.264270 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_aodh-0_33d3a3f3-e9c2-42f3-a13c-34be4404eb1a/aodh-notifier/0.log" Dec 03 19:29:45 crc kubenswrapper[4787]: I1203 19:29:45.431722 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-68677f5b6b-9rwg7_9dc4396d-2f35-42e0-a406-4f0a112fe1cd/barbican-api/0.log" Dec 03 19:29:45 crc kubenswrapper[4787]: I1203 19:29:45.463038 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-68677f5b6b-9rwg7_9dc4396d-2f35-42e0-a406-4f0a112fe1cd/barbican-api-log/0.log" Dec 03 19:29:45 crc kubenswrapper[4787]: I1203 19:29:45.541869 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-78b7c4ccf4-bxfhc_d83ca165-acb3-4e27-b8f3-519897743134/barbican-keystone-listener/0.log" Dec 03 19:29:45 crc kubenswrapper[4787]: I1203 19:29:45.749229 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-78b7c4ccf4-bxfhc_d83ca165-acb3-4e27-b8f3-519897743134/barbican-keystone-listener-log/0.log" Dec 03 19:29:45 crc kubenswrapper[4787]: I1203 19:29:45.812649 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5f6df7d97f-5vmdt_65180fdc-f51b-45e7-ac70-05b9489e1201/barbican-worker-log/0.log" Dec 03 19:29:45 crc kubenswrapper[4787]: I1203 19:29:45.814492 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5f6df7d97f-5vmdt_65180fdc-f51b-45e7-ac70-05b9489e1201/barbican-worker/0.log" Dec 03 19:29:46 crc kubenswrapper[4787]: I1203 19:29:46.022029 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-2fnpx_9490b057-48b6-43c7-ae4f-3d09c1c3e16a/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:46 crc kubenswrapper[4787]: I1203 19:29:46.103176 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_11460db3-3d3a-426a-9980-f1dd41a84497/ceilometer-central-agent/0.log" Dec 03 19:29:46 crc kubenswrapper[4787]: I1203 19:29:46.228113 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_11460db3-3d3a-426a-9980-f1dd41a84497/ceilometer-notification-agent/0.log" Dec 03 19:29:46 crc kubenswrapper[4787]: I1203 19:29:46.248913 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_11460db3-3d3a-426a-9980-f1dd41a84497/proxy-httpd/0.log" Dec 03 19:29:46 crc kubenswrapper[4787]: I1203 19:29:46.321471 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_11460db3-3d3a-426a-9980-f1dd41a84497/sg-core/0.log" Dec 03 19:29:46 crc kubenswrapper[4787]: I1203 19:29:46.424170 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-54wfr_862d8a45-30d2-480c-9aa2-2b8019465b79/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:46 crc kubenswrapper[4787]: I1203 19:29:46.533577 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-247nw_9034ed5d-53f4-4db9-9421-516a5945fce9/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:46 crc kubenswrapper[4787]: I1203 19:29:46.743878 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233/cinder-api/0.log" Dec 03 19:29:46 crc kubenswrapper[4787]: I1203 19:29:46.758223 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_e4a9ddfc-3167-4d7c-bd2a-6ae7a7724233/cinder-api-log/0.log" Dec 03 19:29:46 crc kubenswrapper[4787]: I1203 19:29:46.766696 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:29:46 crc kubenswrapper[4787]: E1203 19:29:46.766958 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:29:47 crc kubenswrapper[4787]: I1203 19:29:47.047637 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_77e12136-eaca-4c6a-9346-da3325061b57/probe/0.log" Dec 03 19:29:47 crc kubenswrapper[4787]: I1203 19:29:47.117069 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_77e12136-eaca-4c6a-9346-da3325061b57/cinder-backup/0.log" Dec 03 19:29:47 crc kubenswrapper[4787]: I1203 19:29:47.158090 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9cb81af5-f082-4509-b5a5-b0a49301b75e/cinder-scheduler/0.log" Dec 03 19:29:47 crc kubenswrapper[4787]: I1203 19:29:47.289101 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_9cb81af5-f082-4509-b5a5-b0a49301b75e/probe/0.log" Dec 03 19:29:47 crc kubenswrapper[4787]: I1203 19:29:47.395089 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_0f2fd27e-97a7-4019-98fd-1ed092285098/cinder-volume/0.log" Dec 03 19:29:47 crc kubenswrapper[4787]: I1203 19:29:47.497465 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_0f2fd27e-97a7-4019-98fd-1ed092285098/probe/0.log" Dec 03 19:29:47 crc kubenswrapper[4787]: I1203 19:29:47.665213 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-4ddrj_6df00b2f-8001-4386-83ed-830a938919e1/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:47 crc kubenswrapper[4787]: I1203 19:29:47.756983 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-vx9s8_d4f19bcf-3f1e-4f60-8098-9812d0dcf2c6/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:48 crc kubenswrapper[4787]: I1203 19:29:48.076142 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-wsddg_060aca68-9071-46af-a627-ed80ad206d74/init/0.log" Dec 03 19:29:48 crc kubenswrapper[4787]: I1203 19:29:48.309744 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-wsddg_060aca68-9071-46af-a627-ed80ad206d74/init/0.log" Dec 03 19:29:48 crc kubenswrapper[4787]: I1203 19:29:48.419565 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_62231118-050c-48ab-9013-0c07ad92cb6b/glance-httpd/0.log" Dec 03 19:29:48 crc kubenswrapper[4787]: I1203 19:29:48.431597 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-5767ddb7c-wsddg_060aca68-9071-46af-a627-ed80ad206d74/dnsmasq-dns/0.log" Dec 03 19:29:48 crc kubenswrapper[4787]: I1203 19:29:48.523366 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_62231118-050c-48ab-9013-0c07ad92cb6b/glance-log/0.log" Dec 03 19:29:48 crc kubenswrapper[4787]: I1203 19:29:48.662436 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_21ef4218-4f36-423e-abd5-86b398276be8/glance-httpd/0.log" Dec 03 19:29:48 crc kubenswrapper[4787]: I1203 19:29:48.694701 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_21ef4218-4f36-423e-abd5-86b398276be8/glance-log/0.log" Dec 03 19:29:49 crc kubenswrapper[4787]: I1203 19:29:49.279763 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-engine-57d98bb984-f2bhk_e9019347-02c2-4e52-86a8-07da730bb3b1/heat-engine/0.log" Dec 03 19:29:49 crc kubenswrapper[4787]: I1203 19:29:49.519419 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5d686989bb-9689v_99c71646-510d-4f03-9308-4d0a9ed3c854/horizon/0.log" Dec 03 19:29:49 crc kubenswrapper[4787]: I1203 19:29:49.768709 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-fzqh7_6d97cc3a-8056-4f1a-98c8-a9dedb9a905d/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:50 crc kubenswrapper[4787]: I1203 19:29:50.020945 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-6z8dd_5f10a9ab-33ee-4938-94f7-870cbdb92c9f/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:50 crc kubenswrapper[4787]: I1203 19:29:50.321242 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-api-74b78cc757-dx9cl_2e92220b-62d7-4d80-8cfa-30a84750f99b/heat-api/0.log" Dec 03 19:29:50 crc kubenswrapper[4787]: I1203 19:29:50.343288 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_heat-cfnapi-796cccd499-c2cvd_ce3bd00e-d1d3-4086-aa4d-d4a48143f533/heat-cfnapi/0.log" Dec 03 19:29:50 crc kubenswrapper[4787]: I1203 19:29:50.382874 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-5d686989bb-9689v_99c71646-510d-4f03-9308-4d0a9ed3c854/horizon-log/0.log" Dec 03 19:29:50 crc kubenswrapper[4787]: I1203 19:29:50.605334 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29413141-r9f29_10d9778c-c2bf-40cf-8934-d584cbc0bad7/keystone-cron/0.log" Dec 03 19:29:50 crc kubenswrapper[4787]: I1203 19:29:50.620753 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29413081-xcmrh_762a8bc2-3aa0-498a-852a-3ede435f23df/keystone-cron/0.log" Dec 03 19:29:50 crc kubenswrapper[4787]: I1203 19:29:50.798606 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-69bdc498c7-wqxdv_c244629d-4ad6-4a6f-9f3d-eaceda01c7e8/keystone-api/0.log" Dec 03 19:29:50 crc kubenswrapper[4787]: I1203 19:29:50.825443 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_c3f8fcb4-9cbd-4029-81a8-8cada9c7b305/kube-state-metrics/0.log" Dec 03 19:29:50 crc kubenswrapper[4787]: I1203 19:29:50.938033 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-jjwfs_7acc721f-ca1b-4a7c-9e11-4c3b58096a51/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:50 crc kubenswrapper[4787]: I1203 19:29:50.997392 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_logging-edpm-deployment-openstack-edpm-ipam-p8gvb_4f182e93-6f85-4e36-978f-50a273b371f0/logging-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:51 crc kubenswrapper[4787]: I1203 19:29:51.112998 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_01cfd003-6a74-4b88-a56d-0daa4cf79daf/manila-api-log/0.log" Dec 03 19:29:51 crc kubenswrapper[4787]: I1203 19:29:51.310440 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_01cfd003-6a74-4b88-a56d-0daa4cf79daf/manila-api/0.log" Dec 03 19:29:51 crc kubenswrapper[4787]: I1203 19:29:51.318399 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_5999e180-294a-4f1a-ae8f-e5a7dbf73b0b/probe/0.log" Dec 03 19:29:51 crc kubenswrapper[4787]: I1203 19:29:51.368551 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_5999e180-294a-4f1a-ae8f-e5a7dbf73b0b/manila-scheduler/0.log" Dec 03 19:29:51 crc kubenswrapper[4787]: I1203 19:29:51.753132 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b52259b1-98e9-4791-a981-fc62fdd0e138/probe/0.log" Dec 03 19:29:51 crc kubenswrapper[4787]: I1203 19:29:51.879480 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_b52259b1-98e9-4791-a981-fc62fdd0e138/manila-share/0.log" Dec 03 19:29:51 crc kubenswrapper[4787]: I1203 19:29:51.979996 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_mysqld-exporter-0_cd6d5b28-0e8a-4ba4-b6f1-716e6f8e4417/mysqld-exporter/0.log" Dec 03 19:29:52 crc kubenswrapper[4787]: I1203 19:29:52.280118 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-gb6mr_424ea689-3407-48b5-a7cc-63bced4b0565/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:52 crc kubenswrapper[4787]: I1203 19:29:52.353252 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9fd5cd74f-xdkkz_593e4809-69d6-43bf-ae94-a5d96431ca2f/neutron-httpd/0.log" Dec 03 19:29:52 crc kubenswrapper[4787]: I1203 19:29:52.401340 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-9fd5cd74f-xdkkz_593e4809-69d6-43bf-ae94-a5d96431ca2f/neutron-api/0.log" Dec 03 19:29:53 crc kubenswrapper[4787]: I1203 19:29:53.167140 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_42612f1a-6592-480c-b475-865267ee2f9f/nova-cell0-conductor-conductor/0.log" Dec 03 19:29:53 crc kubenswrapper[4787]: I1203 19:29:53.412156 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_70d63cec-994c-4594-8521-c9e8b5252363/nova-cell1-conductor-conductor/0.log" Dec 03 19:29:53 crc kubenswrapper[4787]: I1203 19:29:53.474970 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_98746b52-f9a2-4d4f-ab95-2a1c5d7429ec/nova-api-log/0.log" Dec 03 19:29:53 crc kubenswrapper[4787]: I1203 19:29:53.761422 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-2ftbk_4fea1a97-fa55-49e0-ab06-88ec531f4792/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:53 crc kubenswrapper[4787]: I1203 19:29:53.869140 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_47dc3b7d-f87f-4e03-b969-242a9fdbee2b/nova-cell1-novncproxy-novncproxy/0.log" Dec 03 19:29:54 crc kubenswrapper[4787]: I1203 19:29:54.073802 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_be1a822f-d457-417d-9f15-7c3f5b309d7c/nova-metadata-log/0.log" Dec 03 19:29:54 crc kubenswrapper[4787]: I1203 19:29:54.121737 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_98746b52-f9a2-4d4f-ab95-2a1c5d7429ec/nova-api-api/0.log" Dec 03 19:29:54 crc kubenswrapper[4787]: I1203 19:29:54.424299 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d207ad18-746f-4012-a399-2876e6d6bc10/mysql-bootstrap/0.log" Dec 03 19:29:54 crc kubenswrapper[4787]: I1203 19:29:54.720877 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_1a00b714-bec1-411f-a649-c9825253f05e/nova-scheduler-scheduler/0.log" Dec 03 19:29:54 crc kubenswrapper[4787]: I1203 19:29:54.729908 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d207ad18-746f-4012-a399-2876e6d6bc10/mysql-bootstrap/0.log" Dec 03 19:29:54 crc kubenswrapper[4787]: I1203 19:29:54.740186 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_d207ad18-746f-4012-a399-2876e6d6bc10/galera/0.log" Dec 03 19:29:55 crc kubenswrapper[4787]: I1203 19:29:55.002941 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_02a278fe-fdca-423f-a302-59a08cf74ba4/mysql-bootstrap/0.log" Dec 03 19:29:55 crc kubenswrapper[4787]: I1203 19:29:55.383043 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_02a278fe-fdca-423f-a302-59a08cf74ba4/mysql-bootstrap/0.log" Dec 03 19:29:55 crc kubenswrapper[4787]: I1203 19:29:55.478387 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_02a278fe-fdca-423f-a302-59a08cf74ba4/galera/0.log" Dec 03 19:29:55 crc kubenswrapper[4787]: I1203 19:29:55.582451 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_ff54f756-3848-48ec-a235-d2814ff8d7f8/openstackclient/0.log" Dec 03 19:29:55 crc kubenswrapper[4787]: I1203 19:29:55.684048 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-mbdkp_cbb0a5a5-509c-4786-8225-22ed73d9ab20/openstack-network-exporter/0.log" Dec 03 19:29:55 crc kubenswrapper[4787]: I1203 19:29:55.890212 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mc6r6_cab2a5b3-1157-45ab-ad4b-828a1638fd78/ovsdb-server-init/0.log" Dec 03 19:29:56 crc kubenswrapper[4787]: I1203 19:29:56.095932 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mc6r6_cab2a5b3-1157-45ab-ad4b-828a1638fd78/ovsdb-server-init/0.log" Dec 03 19:29:56 crc kubenswrapper[4787]: I1203 19:29:56.134874 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mc6r6_cab2a5b3-1157-45ab-ad4b-828a1638fd78/ovs-vswitchd/0.log" Dec 03 19:29:56 crc kubenswrapper[4787]: I1203 19:29:56.140762 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-mc6r6_cab2a5b3-1157-45ab-ad4b-828a1638fd78/ovsdb-server/0.log" Dec 03 19:29:56 crc kubenswrapper[4787]: I1203 19:29:56.427406 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-prdc5_2db4ba58-5d6c-42ff-b386-7765170d9113/ovn-controller/0.log" Dec 03 19:29:56 crc kubenswrapper[4787]: I1203 19:29:56.636191 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-cwgpt_4fe91a7c-36b7-4528-8061-f2c47bede56d/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:56 crc kubenswrapper[4787]: I1203 19:29:56.682895 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_df89848b-864f-477f-bea0-e34cfcaba0fb/openstack-network-exporter/0.log" Dec 03 19:29:56 crc kubenswrapper[4787]: I1203 19:29:56.904066 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_df89848b-864f-477f-bea0-e34cfcaba0fb/ovn-northd/0.log" Dec 03 19:29:56 crc kubenswrapper[4787]: I1203 19:29:56.911241 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6b71b702-ef57-41ac-8ada-1fe782d5092a/openstack-network-exporter/0.log" Dec 03 19:29:57 crc kubenswrapper[4787]: I1203 19:29:57.099107 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_6b71b702-ef57-41ac-8ada-1fe782d5092a/ovsdbserver-nb/0.log" Dec 03 19:29:57 crc kubenswrapper[4787]: I1203 19:29:57.200893 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cd30e7ed-2464-4857-a563-b02f14871d54/openstack-network-exporter/0.log" Dec 03 19:29:57 crc kubenswrapper[4787]: I1203 19:29:57.337899 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_cd30e7ed-2464-4857-a563-b02f14871d54/ovsdbserver-sb/0.log" Dec 03 19:29:57 crc kubenswrapper[4787]: I1203 19:29:57.661694 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-65d88cc55b-crfnv_ea6b1667-020b-4078-894d-1b944b308802/placement-api/0.log" Dec 03 19:29:57 crc kubenswrapper[4787]: I1203 19:29:57.767353 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:29:57 crc kubenswrapper[4787]: E1203 19:29:57.767621 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:29:57 crc kubenswrapper[4787]: I1203 19:29:57.841035 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-65d88cc55b-crfnv_ea6b1667-020b-4078-894d-1b944b308802/placement-log/0.log" Dec 03 19:29:57 crc kubenswrapper[4787]: I1203 19:29:57.908821 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7e89b75-951a-44b1-9362-fa52750d7450/init-config-reloader/0.log" Dec 03 19:29:58 crc kubenswrapper[4787]: I1203 19:29:58.085849 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_be1a822f-d457-417d-9f15-7c3f5b309d7c/nova-metadata-metadata/0.log" Dec 03 19:29:58 crc kubenswrapper[4787]: I1203 19:29:58.146233 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7e89b75-951a-44b1-9362-fa52750d7450/config-reloader/0.log" Dec 03 19:29:58 crc kubenswrapper[4787]: I1203 19:29:58.176761 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7e89b75-951a-44b1-9362-fa52750d7450/init-config-reloader/0.log" Dec 03 19:29:58 crc kubenswrapper[4787]: I1203 19:29:58.189387 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7e89b75-951a-44b1-9362-fa52750d7450/prometheus/0.log" Dec 03 19:29:58 crc kubenswrapper[4787]: I1203 19:29:58.285405 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_prometheus-metric-storage-0_b7e89b75-951a-44b1-9362-fa52750d7450/thanos-sidecar/0.log" Dec 03 19:29:58 crc kubenswrapper[4787]: I1203 19:29:58.439850 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9b8eef40-19d4-4640-9517-f603e62e646f/setup-container/0.log" Dec 03 19:29:58 crc kubenswrapper[4787]: I1203 19:29:58.603757 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9b8eef40-19d4-4640-9517-f603e62e646f/rabbitmq/0.log" Dec 03 19:29:58 crc kubenswrapper[4787]: I1203 19:29:58.696290 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_9b8eef40-19d4-4640-9517-f603e62e646f/setup-container/0.log" Dec 03 19:29:58 crc kubenswrapper[4787]: I1203 19:29:58.758916 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d774ea58-6127-497a-8948-104e60bfe29b/setup-container/0.log" Dec 03 19:29:58 crc kubenswrapper[4787]: I1203 19:29:58.868213 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d774ea58-6127-497a-8948-104e60bfe29b/setup-container/0.log" Dec 03 19:29:59 crc kubenswrapper[4787]: I1203 19:29:59.278989 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_d774ea58-6127-497a-8948-104e60bfe29b/rabbitmq/0.log" Dec 03 19:29:59 crc kubenswrapper[4787]: I1203 19:29:59.280353 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-jstb6_ce4c2dba-e3f9-4745-adbe-92d44c7bb025/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:59 crc kubenswrapper[4787]: I1203 19:29:59.507291 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-qq4jd_be6c98e3-7b90-462f-ac9b-bf1874f97fc8/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:59 crc kubenswrapper[4787]: I1203 19:29:59.510983 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-pfw4d_8fbd7464-0bdd-4280-87a4-f4b14b82b057/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:29:59 crc kubenswrapper[4787]: I1203 19:29:59.787421 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-g87t4_9dcd8dc0-5dfe-4fa9-a491-ca8ac3a565ac/ssh-known-hosts-edpm-deployment/0.log" Dec 03 19:29:59 crc kubenswrapper[4787]: I1203 19:29:59.975440 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-4fpwj_973215cc-7718-421c-b374-49e64bb9af3e/swift-ring-rebalance/0.log" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.000617 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65b7898d75-fjwnb_cf9c3742-1e29-4f87-bb53-9a60ab3c14ee/proxy-server/0.log" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.228081 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d"] Dec 03 19:30:00 crc kubenswrapper[4787]: E1203 19:30:00.228734 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e42111fe-d231-4173-b7f0-278ead2d6929" containerName="container-00" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.228757 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="e42111fe-d231-4173-b7f0-278ead2d6929" containerName="container-00" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.229207 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="e42111fe-d231-4173-b7f0-278ead2d6929" containerName="container-00" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.231532 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.233703 4787 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.233864 4787 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.238963 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d"] Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.239664 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d8429db-b2d4-424d-b069-36554ab20d76-config-volume\") pod \"collect-profiles-29413170-frk9d\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.239786 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67s7s\" (UniqueName: \"kubernetes.io/projected/2d8429db-b2d4-424d-b069-36554ab20d76-kube-api-access-67s7s\") pod \"collect-profiles-29413170-frk9d\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.239815 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d8429db-b2d4-424d-b069-36554ab20d76-secret-volume\") pod \"collect-profiles-29413170-frk9d\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.252682 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-65b7898d75-fjwnb_cf9c3742-1e29-4f87-bb53-9a60ab3c14ee/proxy-httpd/0.log" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.315313 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/account-reaper/0.log" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.342193 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d8429db-b2d4-424d-b069-36554ab20d76-config-volume\") pod \"collect-profiles-29413170-frk9d\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.342474 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67s7s\" (UniqueName: \"kubernetes.io/projected/2d8429db-b2d4-424d-b069-36554ab20d76-kube-api-access-67s7s\") pod \"collect-profiles-29413170-frk9d\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.342576 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d8429db-b2d4-424d-b069-36554ab20d76-secret-volume\") pod \"collect-profiles-29413170-frk9d\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.344379 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d8429db-b2d4-424d-b069-36554ab20d76-config-volume\") pod \"collect-profiles-29413170-frk9d\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.350310 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d8429db-b2d4-424d-b069-36554ab20d76-secret-volume\") pod \"collect-profiles-29413170-frk9d\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.362461 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67s7s\" (UniqueName: \"kubernetes.io/projected/2d8429db-b2d4-424d-b069-36554ab20d76-kube-api-access-67s7s\") pod \"collect-profiles-29413170-frk9d\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.367877 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/account-auditor/0.log" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.573192 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.639267 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/container-auditor/0.log" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.649395 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/account-server/0.log" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.747233 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/account-replicator/0.log" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.755080 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/container-replicator/0.log" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.884922 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/container-server/0.log" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.917611 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/container-updater/0.log" Dec 03 19:30:00 crc kubenswrapper[4787]: I1203 19:30:00.994095 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/object-expirer/0.log" Dec 03 19:30:01 crc kubenswrapper[4787]: I1203 19:30:01.053591 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/object-auditor/0.log" Dec 03 19:30:01 crc kubenswrapper[4787]: I1203 19:30:01.118629 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d"] Dec 03 19:30:01 crc kubenswrapper[4787]: I1203 19:30:01.158968 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/object-replicator/0.log" Dec 03 19:30:01 crc kubenswrapper[4787]: I1203 19:30:01.193207 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" event={"ID":"2d8429db-b2d4-424d-b069-36554ab20d76","Type":"ContainerStarted","Data":"04a415f4eb0f487fb4e43edfd44bd209d7bc29aaae092b957f333460c9541daa"} Dec 03 19:30:01 crc kubenswrapper[4787]: I1203 19:30:01.214244 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/object-server/0.log" Dec 03 19:30:01 crc kubenswrapper[4787]: I1203 19:30:01.300820 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/object-updater/0.log" Dec 03 19:30:01 crc kubenswrapper[4787]: I1203 19:30:01.330261 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/rsync/0.log" Dec 03 19:30:01 crc kubenswrapper[4787]: I1203 19:30:01.358965 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_106d2d96-ed76-4c59-9577-7086e56f9878/swift-recon-cron/0.log" Dec 03 19:30:01 crc kubenswrapper[4787]: I1203 19:30:01.614146 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-edpm-deployment-openstack-edpm-ipam-c4phf_93b51269-04f3-47d9-b357-9b274a2b3cfe/telemetry-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:30:01 crc kubenswrapper[4787]: I1203 19:30:01.664300 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_telemetry-power-monitoring-edpm-deployment-openstack-edpm-fjjpq_69ab81e9-ff7c-40f4-9f76-6f9e06e4caee/telemetry-power-monitoring-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:30:01 crc kubenswrapper[4787]: I1203 19:30:01.940124 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_dba0eca5-4008-4ce5-b555-d6db725f6466/test-operator-logs-container/0.log" Dec 03 19:30:02 crc kubenswrapper[4787]: I1203 19:30:02.203601 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-wxmx7_48e931e4-9962-4d61-ad24-62a2a5660510/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 03 19:30:02 crc kubenswrapper[4787]: I1203 19:30:02.244278 4787 generic.go:334] "Generic (PLEG): container finished" podID="2d8429db-b2d4-424d-b069-36554ab20d76" containerID="62bef84a4f6b117f4888a3da0b843e04967ec07e1c21b19981a84f75d0acc29f" exitCode=0 Dec 03 19:30:02 crc kubenswrapper[4787]: I1203 19:30:02.244326 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" event={"ID":"2d8429db-b2d4-424d-b069-36554ab20d76","Type":"ContainerDied","Data":"62bef84a4f6b117f4888a3da0b843e04967ec07e1c21b19981a84f75d0acc29f"} Dec 03 19:30:03 crc kubenswrapper[4787]: I1203 19:30:03.098946 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest_4c367f96-011e-4a7c-89e6-53c9ed2d1c90/tempest-tests-tempest-tests-runner/0.log" Dec 03 19:30:03 crc kubenswrapper[4787]: I1203 19:30:03.709389 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:03 crc kubenswrapper[4787]: I1203 19:30:03.742962 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67s7s\" (UniqueName: \"kubernetes.io/projected/2d8429db-b2d4-424d-b069-36554ab20d76-kube-api-access-67s7s\") pod \"2d8429db-b2d4-424d-b069-36554ab20d76\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " Dec 03 19:30:03 crc kubenswrapper[4787]: I1203 19:30:03.743093 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d8429db-b2d4-424d-b069-36554ab20d76-config-volume\") pod \"2d8429db-b2d4-424d-b069-36554ab20d76\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " Dec 03 19:30:03 crc kubenswrapper[4787]: I1203 19:30:03.743177 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d8429db-b2d4-424d-b069-36554ab20d76-secret-volume\") pod \"2d8429db-b2d4-424d-b069-36554ab20d76\" (UID: \"2d8429db-b2d4-424d-b069-36554ab20d76\") " Dec 03 19:30:03 crc kubenswrapper[4787]: I1203 19:30:03.750768 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2d8429db-b2d4-424d-b069-36554ab20d76-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "2d8429db-b2d4-424d-b069-36554ab20d76" (UID: "2d8429db-b2d4-424d-b069-36554ab20d76"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 03 19:30:03 crc kubenswrapper[4787]: I1203 19:30:03.756243 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2d8429db-b2d4-424d-b069-36554ab20d76-config-volume" (OuterVolumeSpecName: "config-volume") pod "2d8429db-b2d4-424d-b069-36554ab20d76" (UID: "2d8429db-b2d4-424d-b069-36554ab20d76"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 03 19:30:03 crc kubenswrapper[4787]: I1203 19:30:03.758557 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2d8429db-b2d4-424d-b069-36554ab20d76-kube-api-access-67s7s" (OuterVolumeSpecName: "kube-api-access-67s7s") pod "2d8429db-b2d4-424d-b069-36554ab20d76" (UID: "2d8429db-b2d4-424d-b069-36554ab20d76"). InnerVolumeSpecName "kube-api-access-67s7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:30:03 crc kubenswrapper[4787]: I1203 19:30:03.854876 4787 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2d8429db-b2d4-424d-b069-36554ab20d76-config-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:30:03 crc kubenswrapper[4787]: I1203 19:30:03.855232 4787 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/2d8429db-b2d4-424d-b069-36554ab20d76-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 03 19:30:03 crc kubenswrapper[4787]: I1203 19:30:03.855245 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67s7s\" (UniqueName: \"kubernetes.io/projected/2d8429db-b2d4-424d-b069-36554ab20d76-kube-api-access-67s7s\") on node \"crc\" DevicePath \"\"" Dec 03 19:30:04 crc kubenswrapper[4787]: I1203 19:30:04.266768 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" event={"ID":"2d8429db-b2d4-424d-b069-36554ab20d76","Type":"ContainerDied","Data":"04a415f4eb0f487fb4e43edfd44bd209d7bc29aaae092b957f333460c9541daa"} Dec 03 19:30:04 crc kubenswrapper[4787]: I1203 19:30:04.266811 4787 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="04a415f4eb0f487fb4e43edfd44bd209d7bc29aaae092b957f333460c9541daa" Dec 03 19:30:04 crc kubenswrapper[4787]: I1203 19:30:04.266865 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29413170-frk9d" Dec 03 19:30:04 crc kubenswrapper[4787]: I1203 19:30:04.804352 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn"] Dec 03 19:30:04 crc kubenswrapper[4787]: I1203 19:30:04.836681 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29413125-27rjn"] Dec 03 19:30:05 crc kubenswrapper[4787]: I1203 19:30:05.780804 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="862bfb78-9edb-409f-bdca-9dcc2f6a3cdc" path="/var/lib/kubelet/pods/862bfb78-9edb-409f-bdca-9dcc2f6a3cdc/volumes" Dec 03 19:30:07 crc kubenswrapper[4787]: E1203 19:30:07.915693 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice/crio-04a415f4eb0f487fb4e43edfd44bd209d7bc29aaae092b957f333460c9541daa\": RecentStats: unable to find data in memory cache]" Dec 03 19:30:12 crc kubenswrapper[4787]: I1203 19:30:12.766243 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:30:12 crc kubenswrapper[4787]: E1203 19:30:12.767116 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:30:18 crc kubenswrapper[4787]: E1203 19:30:18.229143 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice/crio-04a415f4eb0f487fb4e43edfd44bd209d7bc29aaae092b957f333460c9541daa\": RecentStats: unable to find data in memory cache]" Dec 03 19:30:20 crc kubenswrapper[4787]: I1203 19:30:20.477163 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_fe87a100-956b-468e-8aa3-164935a19f5f/memcached/0.log" Dec 03 19:30:25 crc kubenswrapper[4787]: I1203 19:30:25.766504 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:30:25 crc kubenswrapper[4787]: E1203 19:30:25.767367 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:30:28 crc kubenswrapper[4787]: E1203 19:30:28.529699 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice/crio-04a415f4eb0f487fb4e43edfd44bd209d7bc29aaae092b957f333460c9541daa\": RecentStats: unable to find data in memory cache]" Dec 03 19:30:33 crc kubenswrapper[4787]: I1203 19:30:33.466644 4787 scope.go:117] "RemoveContainer" containerID="ad4c6b273d25e555d7adc0ad015f11daa6508e6bf741621e5b1de7c5a0d0f9ce" Dec 03 19:30:35 crc kubenswrapper[4787]: I1203 19:30:35.825994 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-bzjxn_43962cfc-342c-49db-83bf-ccde92708a0b/manager/0.log" Dec 03 19:30:35 crc kubenswrapper[4787]: I1203 19:30:35.855577 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-bzjxn_43962cfc-342c-49db-83bf-ccde92708a0b/kube-rbac-proxy/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.044836 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-fgdm7_6d12a43d-d3da-4b99-b48b-519d660d2527/kube-rbac-proxy/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.098168 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-859b6ccc6-fgdm7_6d12a43d-d3da-4b99-b48b-519d660d2527/manager/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.241617 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-mwpr8_02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23/kube-rbac-proxy/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.278998 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-78b4bc895b-mwpr8_02e8e0d3-fd4d-44a5-a6f0-3e3d265e7d23/manager/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.339766 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/util/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.494955 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/util/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.519960 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/pull/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.543733 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/pull/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.734033 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/pull/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.749271 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/extract/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.750169 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_f48072b2df8bd24560ef910cea51ac6fe223831c47b25abf3de9220decwh9fz_76ec0f29-6ea6-42a5-a4cc-43d5a8b928be/util/0.log" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.766671 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:30:36 crc kubenswrapper[4787]: E1203 19:30:36.767067 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:30:36 crc kubenswrapper[4787]: I1203 19:30:36.963091 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-p425n_355352cc-c3f1-4430-8ff0-55777890e135/kube-rbac-proxy/0.log" Dec 03 19:30:37 crc kubenswrapper[4787]: I1203 19:30:37.018648 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-hr6rk_a6f58c54-821e-4f2b-bce7-1750f64f7ed9/kube-rbac-proxy/0.log" Dec 03 19:30:37 crc kubenswrapper[4787]: I1203 19:30:37.090721 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-77987cd8cd-p425n_355352cc-c3f1-4430-8ff0-55777890e135/manager/0.log" Dec 03 19:30:37 crc kubenswrapper[4787]: I1203 19:30:37.245595 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-cjbzf_2a789aa2-cc4a-42a9-b1c0-9d1a648fd509/kube-rbac-proxy/0.log" Dec 03 19:30:37 crc kubenswrapper[4787]: I1203 19:30:37.265573 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-hr6rk_a6f58c54-821e-4f2b-bce7-1750f64f7ed9/manager/0.log" Dec 03 19:30:37 crc kubenswrapper[4787]: I1203 19:30:37.319699 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-cjbzf_2a789aa2-cc4a-42a9-b1c0-9d1a648fd509/manager/0.log" Dec 03 19:30:37 crc kubenswrapper[4787]: I1203 19:30:37.428606 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-phcn8_b514a6d2-89f9-451c-af24-c0c5a49cdd8d/kube-rbac-proxy/0.log" Dec 03 19:30:37 crc kubenswrapper[4787]: I1203 19:30:37.594244 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-bgmsg_8c4a1981-21a3-49ec-a82d-419f8c080fde/kube-rbac-proxy/0.log" Dec 03 19:30:37 crc kubenswrapper[4787]: I1203 19:30:37.742341 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-6c548fd776-bgmsg_8c4a1981-21a3-49ec-a82d-419f8c080fde/manager/0.log" Dec 03 19:30:37 crc kubenswrapper[4787]: I1203 19:30:37.754930 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-57548d458d-phcn8_b514a6d2-89f9-451c-af24-c0c5a49cdd8d/manager/0.log" Dec 03 19:30:37 crc kubenswrapper[4787]: I1203 19:30:37.832474 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-6txhw_f94890f3-3499-483d-9cda-24a377e982a8/kube-rbac-proxy/0.log" Dec 03 19:30:38 crc kubenswrapper[4787]: I1203 19:30:38.242431 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-cd9jr_5c7f6976-b7e2-453d-ad11-6d0e978cd03d/kube-rbac-proxy/0.log" Dec 03 19:30:38 crc kubenswrapper[4787]: I1203 19:30:38.251616 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-6txhw_f94890f3-3499-483d-9cda-24a377e982a8/manager/0.log" Dec 03 19:30:38 crc kubenswrapper[4787]: I1203 19:30:38.270857 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-7c79b5df47-cd9jr_5c7f6976-b7e2-453d-ad11-6d0e978cd03d/manager/0.log" Dec 03 19:30:38 crc kubenswrapper[4787]: I1203 19:30:38.440786 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-5m5h7_6384c305-657a-465d-8e48-487fb12cd52f/kube-rbac-proxy/0.log" Dec 03 19:30:38 crc kubenswrapper[4787]: I1203 19:30:38.492001 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-56bbcc9d85-5m5h7_6384c305-657a-465d-8e48-487fb12cd52f/manager/0.log" Dec 03 19:30:38 crc kubenswrapper[4787]: I1203 19:30:38.700486 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-sbcbk_04fbeb64-47f2-4393-8295-0ac7d6db4df3/manager/0.log" Dec 03 19:30:38 crc kubenswrapper[4787]: I1203 19:30:38.703373 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zwmxn_a0326a22-7690-48ed-84f5-01cc9e8331b2/kube-rbac-proxy/0.log" Dec 03 19:30:38 crc kubenswrapper[4787]: I1203 19:30:38.753368 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-sbcbk_04fbeb64-47f2-4393-8295-0ac7d6db4df3/kube-rbac-proxy/0.log" Dec 03 19:30:38 crc kubenswrapper[4787]: E1203 19:30:38.807442 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice/crio-04a415f4eb0f487fb4e43edfd44bd209d7bc29aaae092b957f333460c9541daa\": RecentStats: unable to find data in memory cache]" Dec 03 19:30:38 crc kubenswrapper[4787]: I1203 19:30:38.989055 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-75r2j_67472bf9-12b8-4463-bc95-dec19e689f36/kube-rbac-proxy/0.log" Dec 03 19:30:39 crc kubenswrapper[4787]: I1203 19:30:39.011315 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-zwmxn_a0326a22-7690-48ed-84f5-01cc9e8331b2/manager/0.log" Dec 03 19:30:39 crc kubenswrapper[4787]: I1203 19:30:39.101195 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-75r2j_67472bf9-12b8-4463-bc95-dec19e689f36/manager/0.log" Dec 03 19:30:39 crc kubenswrapper[4787]: I1203 19:30:39.236391 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz_31c5513f-61f7-4d6c-8136-b212cc171732/kube-rbac-proxy/0.log" Dec 03 19:30:39 crc kubenswrapper[4787]: I1203 19:30:39.277230 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-64bc77cfd4pqrjz_31c5513f-61f7-4d6c-8136-b212cc171732/manager/0.log" Dec 03 19:30:39 crc kubenswrapper[4787]: I1203 19:30:39.825497 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-t5bxl_e5effab2-18ac-4ad1-8d08-618f20509a85/registry-server/0.log" Dec 03 19:30:39 crc kubenswrapper[4787]: I1203 19:30:39.862721 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-5db84b8f89-tm758_3b816649-4a0d-4090-9e9f-df5c5a296777/operator/0.log" Dec 03 19:30:39 crc kubenswrapper[4787]: I1203 19:30:39.922128 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-xg7kr_226cdcb0-0abc-4a06-9d5c-1d30d9ac8288/kube-rbac-proxy/0.log" Dec 03 19:30:40 crc kubenswrapper[4787]: I1203 19:30:40.153489 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-xg7kr_226cdcb0-0abc-4a06-9d5c-1d30d9ac8288/manager/0.log" Dec 03 19:30:40 crc kubenswrapper[4787]: I1203 19:30:40.190646 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-zgkn8_dd3a3971-adb9-46cb-a0b7-63897e326341/kube-rbac-proxy/0.log" Dec 03 19:30:40 crc kubenswrapper[4787]: I1203 19:30:40.356158 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-zgkn8_dd3a3971-adb9-46cb-a0b7-63897e326341/manager/0.log" Dec 03 19:30:40 crc kubenswrapper[4787]: I1203 19:30:40.471435 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-cnjbp_c7e741e1-b6d3-4ceb-a77f-6b648aa5aad1/operator/0.log" Dec 03 19:30:40 crc kubenswrapper[4787]: I1203 19:30:40.589297 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-vmpfz_c209ad3f-9014-4385-96cb-323dd27b8335/kube-rbac-proxy/0.log" Dec 03 19:30:40 crc kubenswrapper[4787]: I1203 19:30:40.685900 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f8c65bbfc-vmpfz_c209ad3f-9014-4385-96cb-323dd27b8335/manager/0.log" Dec 03 19:30:40 crc kubenswrapper[4787]: I1203 19:30:40.809195 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-66f6f55998-k4kg2_a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173/kube-rbac-proxy/0.log" Dec 03 19:30:40 crc kubenswrapper[4787]: I1203 19:30:40.907620 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wtvd4_d1364502-c591-47c8-b24c-00c77e3a8d23/kube-rbac-proxy/0.log" Dec 03 19:30:41 crc kubenswrapper[4787]: I1203 19:30:41.076542 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-5854674fcc-wtvd4_d1364502-c591-47c8-b24c-00c77e3a8d23/manager/0.log" Dec 03 19:30:41 crc kubenswrapper[4787]: I1203 19:30:41.175532 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-5574b8467d-tg2v7_d9b27877-9ba0-4487-a487-aceafca2075a/manager/0.log" Dec 03 19:30:41 crc kubenswrapper[4787]: I1203 19:30:41.201101 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-66f6f55998-k4kg2_a2e5ad5b-87b4-4a79-a23f-dccbd7bcc173/manager/0.log" Dec 03 19:30:41 crc kubenswrapper[4787]: I1203 19:30:41.227128 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-2wq6s_8af08022-68dd-4fc4-a052-2cfde5f1ab0a/kube-rbac-proxy/0.log" Dec 03 19:30:41 crc kubenswrapper[4787]: I1203 19:30:41.304669 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-769dc69bc-2wq6s_8af08022-68dd-4fc4-a052-2cfde5f1ab0a/manager/0.log" Dec 03 19:30:48 crc kubenswrapper[4787]: I1203 19:30:48.766908 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:30:48 crc kubenswrapper[4787]: E1203 19:30:48.767696 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:30:49 crc kubenswrapper[4787]: E1203 19:30:49.143785 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice/crio-04a415f4eb0f487fb4e43edfd44bd209d7bc29aaae092b957f333460c9541daa\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice\": RecentStats: unable to find data in memory cache]" Dec 03 19:30:59 crc kubenswrapper[4787]: E1203 19:30:59.480980 4787 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d8429db_b2d4_424d_b069_36554ab20d76.slice/crio-04a415f4eb0f487fb4e43edfd44bd209d7bc29aaae092b957f333460c9541daa\": RecentStats: unable to find data in memory cache]" Dec 03 19:30:59 crc kubenswrapper[4787]: I1203 19:30:59.766367 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:30:59 crc kubenswrapper[4787]: E1203 19:30:59.766715 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:31:02 crc kubenswrapper[4787]: I1203 19:31:02.403040 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bj8b6_ee54b630-23ff-4200-aa84-d3aca72f50e8/control-plane-machine-set-operator/0.log" Dec 03 19:31:02 crc kubenswrapper[4787]: I1203 19:31:02.588741 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sm94s_48860c61-124e-43cf-9cf6-fd36f33866f5/kube-rbac-proxy/0.log" Dec 03 19:31:02 crc kubenswrapper[4787]: I1203 19:31:02.620214 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-sm94s_48860c61-124e-43cf-9cf6-fd36f33866f5/machine-api-operator/0.log" Dec 03 19:31:10 crc kubenswrapper[4787]: I1203 19:31:10.768302 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:31:10 crc kubenswrapper[4787]: E1203 19:31:10.769147 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:31:16 crc kubenswrapper[4787]: I1203 19:31:16.278605 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-hrd48_249fd549-e9f6-4a50-9b84-b96e46f2791f/cert-manager-controller/0.log" Dec 03 19:31:16 crc kubenswrapper[4787]: I1203 19:31:16.432442 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-r8mcm_c0bb5afc-e67f-42c5-849e-2892c2bf7d3a/cert-manager-cainjector/0.log" Dec 03 19:31:16 crc kubenswrapper[4787]: I1203 19:31:16.475958 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-f46ld_f5b1552f-1b75-4000-b31e-4d247ea51a65/cert-manager-webhook/0.log" Dec 03 19:31:25 crc kubenswrapper[4787]: I1203 19:31:25.765452 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:31:25 crc kubenswrapper[4787]: E1203 19:31:25.766238 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:31:29 crc kubenswrapper[4787]: I1203 19:31:29.395161 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-j2lgj_ce1a6a45-6907-432d-a9c6-3d1f1b9fad52/nmstate-console-plugin/0.log" Dec 03 19:31:29 crc kubenswrapper[4787]: I1203 19:31:29.500988 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-njhbf_f8584654-d601-42cc-98e4-1abd6fdbe848/nmstate-handler/0.log" Dec 03 19:31:29 crc kubenswrapper[4787]: I1203 19:31:29.577377 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-p9brp_2a021e2e-e7b3-44ed-8bc2-3008461b97b3/nmstate-metrics/0.log" Dec 03 19:31:29 crc kubenswrapper[4787]: I1203 19:31:29.582231 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-p9brp_2a021e2e-e7b3-44ed-8bc2-3008461b97b3/kube-rbac-proxy/0.log" Dec 03 19:31:29 crc kubenswrapper[4787]: I1203 19:31:29.791905 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-b4gp4_2915592e-3cf5-43c3-a142-d89b98274df2/nmstate-webhook/0.log" Dec 03 19:31:29 crc kubenswrapper[4787]: I1203 19:31:29.794279 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-fhrgj_e4096b9f-51a5-4bb4-860e-0f689a2d9f07/nmstate-operator/0.log" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.025082 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-ntrzw"] Dec 03 19:31:34 crc kubenswrapper[4787]: E1203 19:31:34.026344 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2d8429db-b2d4-424d-b069-36554ab20d76" containerName="collect-profiles" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.026358 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="2d8429db-b2d4-424d-b069-36554ab20d76" containerName="collect-profiles" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.026638 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="2d8429db-b2d4-424d-b069-36554ab20d76" containerName="collect-profiles" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.028272 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.037872 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntrzw"] Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.094573 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-catalog-content\") pod \"redhat-marketplace-ntrzw\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.095192 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-utilities\") pod \"redhat-marketplace-ntrzw\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.095744 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ms27j\" (UniqueName: \"kubernetes.io/projected/f17e1369-d28b-4930-8d43-6f17d9289a7d-kube-api-access-ms27j\") pod \"redhat-marketplace-ntrzw\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.198436 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ms27j\" (UniqueName: \"kubernetes.io/projected/f17e1369-d28b-4930-8d43-6f17d9289a7d-kube-api-access-ms27j\") pod \"redhat-marketplace-ntrzw\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.198539 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-catalog-content\") pod \"redhat-marketplace-ntrzw\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.198606 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-utilities\") pod \"redhat-marketplace-ntrzw\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.199219 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-utilities\") pod \"redhat-marketplace-ntrzw\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.199266 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-catalog-content\") pod \"redhat-marketplace-ntrzw\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.222878 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ms27j\" (UniqueName: \"kubernetes.io/projected/f17e1369-d28b-4930-8d43-6f17d9289a7d-kube-api-access-ms27j\") pod \"redhat-marketplace-ntrzw\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.361732 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:34 crc kubenswrapper[4787]: I1203 19:31:34.869653 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntrzw"] Dec 03 19:31:35 crc kubenswrapper[4787]: I1203 19:31:35.200240 4787 generic.go:334] "Generic (PLEG): container finished" podID="f17e1369-d28b-4930-8d43-6f17d9289a7d" containerID="b8332845555974c5451c4f4e3afa66a8ad5bf9d38396af8e67ec762b1bf21fc5" exitCode=0 Dec 03 19:31:35 crc kubenswrapper[4787]: I1203 19:31:35.200355 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntrzw" event={"ID":"f17e1369-d28b-4930-8d43-6f17d9289a7d","Type":"ContainerDied","Data":"b8332845555974c5451c4f4e3afa66a8ad5bf9d38396af8e67ec762b1bf21fc5"} Dec 03 19:31:35 crc kubenswrapper[4787]: I1203 19:31:35.201179 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntrzw" event={"ID":"f17e1369-d28b-4930-8d43-6f17d9289a7d","Type":"ContainerStarted","Data":"c07e58231c6857bb4fb5ba44858344152134cf8101f75581fb1e7f4095e5cd41"} Dec 03 19:31:36 crc kubenswrapper[4787]: I1203 19:31:36.766783 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:31:36 crc kubenswrapper[4787]: E1203 19:31:36.767683 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:31:37 crc kubenswrapper[4787]: I1203 19:31:37.228425 4787 generic.go:334] "Generic (PLEG): container finished" podID="f17e1369-d28b-4930-8d43-6f17d9289a7d" containerID="500a851de09885ed768b54bef39039967588ce1f59c86d7b7179ee4fc2f6b2c3" exitCode=0 Dec 03 19:31:37 crc kubenswrapper[4787]: I1203 19:31:37.228508 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntrzw" event={"ID":"f17e1369-d28b-4930-8d43-6f17d9289a7d","Type":"ContainerDied","Data":"500a851de09885ed768b54bef39039967588ce1f59c86d7b7179ee4fc2f6b2c3"} Dec 03 19:31:38 crc kubenswrapper[4787]: I1203 19:31:38.253485 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntrzw" event={"ID":"f17e1369-d28b-4930-8d43-6f17d9289a7d","Type":"ContainerStarted","Data":"722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109"} Dec 03 19:31:38 crc kubenswrapper[4787]: I1203 19:31:38.284133 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-ntrzw" podStartSLOduration=2.794458081 podStartE2EDuration="5.284103261s" podCreationTimestamp="2025-12-03 19:31:33 +0000 UTC" firstStartedPulling="2025-12-03 19:31:35.202107664 +0000 UTC m=+8312.019578623" lastFinishedPulling="2025-12-03 19:31:37.691752834 +0000 UTC m=+8314.509223803" observedRunningTime="2025-12-03 19:31:38.27573326 +0000 UTC m=+8315.093204219" watchObservedRunningTime="2025-12-03 19:31:38.284103261 +0000 UTC m=+8315.101574220" Dec 03 19:31:43 crc kubenswrapper[4787]: I1203 19:31:43.887590 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5cd867f87d-zwxll_c73f5b31-0cde-406e-800b-153981eeead6/kube-rbac-proxy/0.log" Dec 03 19:31:43 crc kubenswrapper[4787]: I1203 19:31:43.937522 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5cd867f87d-zwxll_c73f5b31-0cde-406e-800b-153981eeead6/manager/0.log" Dec 03 19:31:44 crc kubenswrapper[4787]: I1203 19:31:44.362644 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:44 crc kubenswrapper[4787]: I1203 19:31:44.362691 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:44 crc kubenswrapper[4787]: I1203 19:31:44.415480 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:45 crc kubenswrapper[4787]: I1203 19:31:45.370739 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:45 crc kubenswrapper[4787]: I1203 19:31:45.428438 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntrzw"] Dec 03 19:31:47 crc kubenswrapper[4787]: I1203 19:31:47.339110 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-ntrzw" podUID="f17e1369-d28b-4930-8d43-6f17d9289a7d" containerName="registry-server" containerID="cri-o://722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109" gracePeriod=2 Dec 03 19:31:47 crc kubenswrapper[4787]: I1203 19:31:47.932173 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.036266 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-utilities\") pod \"f17e1369-d28b-4930-8d43-6f17d9289a7d\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.036617 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-catalog-content\") pod \"f17e1369-d28b-4930-8d43-6f17d9289a7d\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.036676 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ms27j\" (UniqueName: \"kubernetes.io/projected/f17e1369-d28b-4930-8d43-6f17d9289a7d-kube-api-access-ms27j\") pod \"f17e1369-d28b-4930-8d43-6f17d9289a7d\" (UID: \"f17e1369-d28b-4930-8d43-6f17d9289a7d\") " Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.037006 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-utilities" (OuterVolumeSpecName: "utilities") pod "f17e1369-d28b-4930-8d43-6f17d9289a7d" (UID: "f17e1369-d28b-4930-8d43-6f17d9289a7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.038570 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.042424 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f17e1369-d28b-4930-8d43-6f17d9289a7d-kube-api-access-ms27j" (OuterVolumeSpecName: "kube-api-access-ms27j") pod "f17e1369-d28b-4930-8d43-6f17d9289a7d" (UID: "f17e1369-d28b-4930-8d43-6f17d9289a7d"). InnerVolumeSpecName "kube-api-access-ms27j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.054549 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f17e1369-d28b-4930-8d43-6f17d9289a7d" (UID: "f17e1369-d28b-4930-8d43-6f17d9289a7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.144385 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f17e1369-d28b-4930-8d43-6f17d9289a7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.144424 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ms27j\" (UniqueName: \"kubernetes.io/projected/f17e1369-d28b-4930-8d43-6f17d9289a7d-kube-api-access-ms27j\") on node \"crc\" DevicePath \"\"" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.357891 4787 generic.go:334] "Generic (PLEG): container finished" podID="f17e1369-d28b-4930-8d43-6f17d9289a7d" containerID="722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109" exitCode=0 Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.357939 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntrzw" event={"ID":"f17e1369-d28b-4930-8d43-6f17d9289a7d","Type":"ContainerDied","Data":"722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109"} Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.357972 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-ntrzw" event={"ID":"f17e1369-d28b-4930-8d43-6f17d9289a7d","Type":"ContainerDied","Data":"c07e58231c6857bb4fb5ba44858344152134cf8101f75581fb1e7f4095e5cd41"} Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.357996 4787 scope.go:117] "RemoveContainer" containerID="722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.358043 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-ntrzw" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.398953 4787 scope.go:117] "RemoveContainer" containerID="500a851de09885ed768b54bef39039967588ce1f59c86d7b7179ee4fc2f6b2c3" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.405449 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntrzw"] Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.419269 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-ntrzw"] Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.431097 4787 scope.go:117] "RemoveContainer" containerID="b8332845555974c5451c4f4e3afa66a8ad5bf9d38396af8e67ec762b1bf21fc5" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.498392 4787 scope.go:117] "RemoveContainer" containerID="722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109" Dec 03 19:31:48 crc kubenswrapper[4787]: E1203 19:31:48.499010 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109\": container with ID starting with 722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109 not found: ID does not exist" containerID="722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.499102 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109"} err="failed to get container status \"722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109\": rpc error: code = NotFound desc = could not find container \"722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109\": container with ID starting with 722e0904f9bb771b8bb139071a4440201bb4f75d68053bf229e7c97ba08c0109 not found: ID does not exist" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.499143 4787 scope.go:117] "RemoveContainer" containerID="500a851de09885ed768b54bef39039967588ce1f59c86d7b7179ee4fc2f6b2c3" Dec 03 19:31:48 crc kubenswrapper[4787]: E1203 19:31:48.499618 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"500a851de09885ed768b54bef39039967588ce1f59c86d7b7179ee4fc2f6b2c3\": container with ID starting with 500a851de09885ed768b54bef39039967588ce1f59c86d7b7179ee4fc2f6b2c3 not found: ID does not exist" containerID="500a851de09885ed768b54bef39039967588ce1f59c86d7b7179ee4fc2f6b2c3" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.499673 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"500a851de09885ed768b54bef39039967588ce1f59c86d7b7179ee4fc2f6b2c3"} err="failed to get container status \"500a851de09885ed768b54bef39039967588ce1f59c86d7b7179ee4fc2f6b2c3\": rpc error: code = NotFound desc = could not find container \"500a851de09885ed768b54bef39039967588ce1f59c86d7b7179ee4fc2f6b2c3\": container with ID starting with 500a851de09885ed768b54bef39039967588ce1f59c86d7b7179ee4fc2f6b2c3 not found: ID does not exist" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.499696 4787 scope.go:117] "RemoveContainer" containerID="b8332845555974c5451c4f4e3afa66a8ad5bf9d38396af8e67ec762b1bf21fc5" Dec 03 19:31:48 crc kubenswrapper[4787]: E1203 19:31:48.500072 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8332845555974c5451c4f4e3afa66a8ad5bf9d38396af8e67ec762b1bf21fc5\": container with ID starting with b8332845555974c5451c4f4e3afa66a8ad5bf9d38396af8e67ec762b1bf21fc5 not found: ID does not exist" containerID="b8332845555974c5451c4f4e3afa66a8ad5bf9d38396af8e67ec762b1bf21fc5" Dec 03 19:31:48 crc kubenswrapper[4787]: I1203 19:31:48.500118 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8332845555974c5451c4f4e3afa66a8ad5bf9d38396af8e67ec762b1bf21fc5"} err="failed to get container status \"b8332845555974c5451c4f4e3afa66a8ad5bf9d38396af8e67ec762b1bf21fc5\": rpc error: code = NotFound desc = could not find container \"b8332845555974c5451c4f4e3afa66a8ad5bf9d38396af8e67ec762b1bf21fc5\": container with ID starting with b8332845555974c5451c4f4e3afa66a8ad5bf9d38396af8e67ec762b1bf21fc5 not found: ID does not exist" Dec 03 19:31:49 crc kubenswrapper[4787]: I1203 19:31:49.781894 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f17e1369-d28b-4930-8d43-6f17d9289a7d" path="/var/lib/kubelet/pods/f17e1369-d28b-4930-8d43-6f17d9289a7d/volumes" Dec 03 19:31:51 crc kubenswrapper[4787]: I1203 19:31:51.766407 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:31:51 crc kubenswrapper[4787]: E1203 19:31:51.767537 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:31:58 crc kubenswrapper[4787]: I1203 19:31:58.832012 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_cluster-logging-operator-ff9846bd-z7f2j_28f7de1d-ea99-4b90-bdb9-bc1f04f201ff/cluster-logging-operator/0.log" Dec 03 19:31:58 crc kubenswrapper[4787]: I1203 19:31:58.991295 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_collector-ghzsk_44f70e1d-0f76-4acb-bf6b-1a5ca89469ca/collector/0.log" Dec 03 19:31:59 crc kubenswrapper[4787]: I1203 19:31:59.015064 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-compactor-0_db4243ab-4a52-4991-89a6-96e1ef9a348f/loki-compactor/0.log" Dec 03 19:31:59 crc kubenswrapper[4787]: I1203 19:31:59.194345 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-distributor-76cc67bf56-flghj_e1be9303-41d5-431a-bb49-b0e104ce4625/loki-distributor/0.log" Dec 03 19:31:59 crc kubenswrapper[4787]: I1203 19:31:59.205488 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-57f76f4c5d-dzw77_35486d7a-dcc2-40bf-b2b0-869ecfe46ccc/gateway/0.log" Dec 03 19:31:59 crc kubenswrapper[4787]: I1203 19:31:59.285188 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-57f76f4c5d-dzw77_35486d7a-dcc2-40bf-b2b0-869ecfe46ccc/opa/0.log" Dec 03 19:31:59 crc kubenswrapper[4787]: I1203 19:31:59.377585 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-57f76f4c5d-v45xd_cb532cad-464c-4fe6-99de-b7897de9ac51/gateway/0.log" Dec 03 19:31:59 crc kubenswrapper[4787]: I1203 19:31:59.411742 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-gateway-57f76f4c5d-v45xd_cb532cad-464c-4fe6-99de-b7897de9ac51/opa/0.log" Dec 03 19:31:59 crc kubenswrapper[4787]: I1203 19:31:59.571182 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-index-gateway-0_b4de518a-6a4d-441b-b04c-d35ca9a8b5f9/loki-index-gateway/0.log" Dec 03 19:31:59 crc kubenswrapper[4787]: I1203 19:31:59.744783 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-ingester-0_7da69ee7-38ad-4cce-9bf0-140e7283a775/loki-ingester/0.log" Dec 03 19:31:59 crc kubenswrapper[4787]: I1203 19:31:59.820504 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-querier-5895d59bb8-xq92s_4e0f3ec8-7217-4cc2-ad4b-7f5f54213f1a/loki-querier/0.log" Dec 03 19:31:59 crc kubenswrapper[4787]: I1203 19:31:59.936234 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-logging_logging-loki-query-frontend-84558f7c9f-zsjlj_b43db9b8-11f8-4fb7-82d2-6b38e34e1a02/loki-query-frontend/0.log" Dec 03 19:32:06 crc kubenswrapper[4787]: I1203 19:32:06.766603 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:32:06 crc kubenswrapper[4787]: E1203 19:32:06.767407 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:32:14 crc kubenswrapper[4787]: I1203 19:32:14.975527 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6rp2q_e1c08eda-0859-42f8-9ca2-f1c4df77038c/kube-rbac-proxy/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.146315 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-6rp2q_e1c08eda-0859-42f8-9ca2-f1c4df77038c/controller/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.190600 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-frr-files/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.431544 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-reloader/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.432085 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-frr-files/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.441749 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-metrics/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.505818 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-reloader/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.691853 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-frr-files/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.747281 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-reloader/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.772833 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-metrics/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.782871 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-metrics/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.909307 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-frr-files/0.log" Dec 03 19:32:15 crc kubenswrapper[4787]: I1203 19:32:15.941278 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-reloader/0.log" Dec 03 19:32:16 crc kubenswrapper[4787]: I1203 19:32:16.097478 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/cp-metrics/0.log" Dec 03 19:32:16 crc kubenswrapper[4787]: I1203 19:32:16.211814 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/controller/0.log" Dec 03 19:32:16 crc kubenswrapper[4787]: I1203 19:32:16.445435 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/kube-rbac-proxy/0.log" Dec 03 19:32:16 crc kubenswrapper[4787]: I1203 19:32:16.507043 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/frr-metrics/0.log" Dec 03 19:32:16 crc kubenswrapper[4787]: I1203 19:32:16.570094 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/kube-rbac-proxy-frr/0.log" Dec 03 19:32:16 crc kubenswrapper[4787]: I1203 19:32:16.697857 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/reloader/0.log" Dec 03 19:32:16 crc kubenswrapper[4787]: I1203 19:32:16.825540 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-kqsb2_55b2904e-d04c-43ef-b054-0ef8636a2316/frr-k8s-webhook-server/0.log" Dec 03 19:32:17 crc kubenswrapper[4787]: I1203 19:32:17.759878 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-9b64dcbdc-qzmjb_fe122b60-4167-40b6-8562-1b52112f44f8/manager/0.log" Dec 03 19:32:18 crc kubenswrapper[4787]: I1203 19:32:18.104777 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-86b8bcb77d-fd8v5_200c13aa-bae6-4b29-b6b1-35b625ce2dfc/webhook-server/0.log" Dec 03 19:32:18 crc kubenswrapper[4787]: I1203 19:32:18.202256 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sfcqc_d571a3dc-bd38-4881-857f-e7986b1d90af/kube-rbac-proxy/0.log" Dec 03 19:32:18 crc kubenswrapper[4787]: I1203 19:32:18.447596 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-pgt2s_37601185-c3bd-4614-b119-05f1b07f2875/frr/0.log" Dec 03 19:32:18 crc kubenswrapper[4787]: I1203 19:32:18.842617 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-sfcqc_d571a3dc-bd38-4881-857f-e7986b1d90af/speaker/0.log" Dec 03 19:32:21 crc kubenswrapper[4787]: I1203 19:32:21.766151 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:32:21 crc kubenswrapper[4787]: E1203 19:32:21.767086 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:32:33 crc kubenswrapper[4787]: I1203 19:32:33.777270 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:32:33 crc kubenswrapper[4787]: E1203 19:32:33.778247 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:32:34 crc kubenswrapper[4787]: I1203 19:32:34.264584 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/util/0.log" Dec 03 19:32:34 crc kubenswrapper[4787]: I1203 19:32:34.490734 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/util/0.log" Dec 03 19:32:34 crc kubenswrapper[4787]: I1203 19:32:34.531695 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/pull/0.log" Dec 03 19:32:34 crc kubenswrapper[4787]: I1203 19:32:34.557505 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/pull/0.log" Dec 03 19:32:34 crc kubenswrapper[4787]: I1203 19:32:34.786670 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/util/0.log" Dec 03 19:32:34 crc kubenswrapper[4787]: I1203 19:32:34.802522 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/pull/0.log" Dec 03 19:32:34 crc kubenswrapper[4787]: I1203 19:32:34.877134 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_4529ed37fc81381df2b45ea09e6f1b4af8d1558d603912431befd8aeb8p8x4j_74207f9d-ab04-4834-9813-5d89fe068d43/extract/0.log" Dec 03 19:32:35 crc kubenswrapper[4787]: I1203 19:32:35.056629 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/util/0.log" Dec 03 19:32:35 crc kubenswrapper[4787]: I1203 19:32:35.215669 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/pull/0.log" Dec 03 19:32:35 crc kubenswrapper[4787]: I1203 19:32:35.258138 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/util/0.log" Dec 03 19:32:35 crc kubenswrapper[4787]: I1203 19:32:35.281497 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/pull/0.log" Dec 03 19:32:35 crc kubenswrapper[4787]: I1203 19:32:35.447853 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/util/0.log" Dec 03 19:32:35 crc kubenswrapper[4787]: I1203 19:32:35.466476 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/pull/0.log" Dec 03 19:32:35 crc kubenswrapper[4787]: I1203 19:32:35.507728 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fkwss8_7a1e8379-9aab-4e95-8c05-089edfed5d4b/extract/0.log" Dec 03 19:32:35 crc kubenswrapper[4787]: I1203 19:32:35.641765 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/util/0.log" Dec 03 19:32:35 crc kubenswrapper[4787]: I1203 19:32:35.801777 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/pull/0.log" Dec 03 19:32:35 crc kubenswrapper[4787]: I1203 19:32:35.833984 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/pull/0.log" Dec 03 19:32:35 crc kubenswrapper[4787]: I1203 19:32:35.845792 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/util/0.log" Dec 03 19:32:36 crc kubenswrapper[4787]: I1203 19:32:36.010536 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/pull/0.log" Dec 03 19:32:36 crc kubenswrapper[4787]: I1203 19:32:36.038579 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/util/0.log" Dec 03 19:32:36 crc kubenswrapper[4787]: I1203 19:32:36.081682 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6c372a8d094fad7255d3bbeabb4914bd2356af7b203a2d2176be1c9210tglvc_e3390472-0d01-4c86-8756-cca8bc29a289/extract/0.log" Dec 03 19:32:36 crc kubenswrapper[4787]: I1203 19:32:36.251583 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/util/0.log" Dec 03 19:32:36 crc kubenswrapper[4787]: I1203 19:32:36.437196 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/util/0.log" Dec 03 19:32:36 crc kubenswrapper[4787]: I1203 19:32:36.463730 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/pull/0.log" Dec 03 19:32:36 crc kubenswrapper[4787]: I1203 19:32:36.467825 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/pull/0.log" Dec 03 19:32:36 crc kubenswrapper[4787]: I1203 19:32:36.642512 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/util/0.log" Dec 03 19:32:36 crc kubenswrapper[4787]: I1203 19:32:36.696765 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/extract/0.log" Dec 03 19:32:36 crc kubenswrapper[4787]: I1203 19:32:36.703182 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_a8a03f72555e3294619fd3c0a789fa82d1f6921a8cf9935ed9b211463fsgl6j_9567b7ba-009e-4ebb-a03f-abd6ed1e9aa0/pull/0.log" Dec 03 19:32:36 crc kubenswrapper[4787]: I1203 19:32:36.844774 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/util/0.log" Dec 03 19:32:37 crc kubenswrapper[4787]: I1203 19:32:37.067583 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/util/0.log" Dec 03 19:32:37 crc kubenswrapper[4787]: I1203 19:32:37.115008 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/pull/0.log" Dec 03 19:32:37 crc kubenswrapper[4787]: I1203 19:32:37.135979 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/pull/0.log" Dec 03 19:32:37 crc kubenswrapper[4787]: I1203 19:32:37.378139 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/util/0.log" Dec 03 19:32:37 crc kubenswrapper[4787]: I1203 19:32:37.438826 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/extract/0.log" Dec 03 19:32:37 crc kubenswrapper[4787]: I1203 19:32:37.467200 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lr5d8_7bed6e06-e6f3-4611-90fd-8963706e05d5/pull/0.log" Dec 03 19:32:37 crc kubenswrapper[4787]: I1203 19:32:37.864323 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-utilities/0.log" Dec 03 19:32:38 crc kubenswrapper[4787]: I1203 19:32:38.137591 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-utilities/0.log" Dec 03 19:32:38 crc kubenswrapper[4787]: I1203 19:32:38.152878 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-content/0.log" Dec 03 19:32:38 crc kubenswrapper[4787]: I1203 19:32:38.166170 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-content/0.log" Dec 03 19:32:38 crc kubenswrapper[4787]: I1203 19:32:38.341517 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-content/0.log" Dec 03 19:32:38 crc kubenswrapper[4787]: I1203 19:32:38.358670 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/extract-utilities/0.log" Dec 03 19:32:38 crc kubenswrapper[4787]: I1203 19:32:38.425892 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-utilities/0.log" Dec 03 19:32:38 crc kubenswrapper[4787]: I1203 19:32:38.741977 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-content/0.log" Dec 03 19:32:38 crc kubenswrapper[4787]: I1203 19:32:38.762385 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-utilities/0.log" Dec 03 19:32:38 crc kubenswrapper[4787]: I1203 19:32:38.827567 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-content/0.log" Dec 03 19:32:38 crc kubenswrapper[4787]: I1203 19:32:38.970547 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-content/0.log" Dec 03 19:32:39 crc kubenswrapper[4787]: I1203 19:32:39.039993 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/extract-utilities/0.log" Dec 03 19:32:39 crc kubenswrapper[4787]: I1203 19:32:39.252073 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-278bm_38b06bc8-6c75-46f8-a53b-d95109183306/marketplace-operator/0.log" Dec 03 19:32:39 crc kubenswrapper[4787]: I1203 19:32:39.452151 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-utilities/0.log" Dec 03 19:32:39 crc kubenswrapper[4787]: I1203 19:32:39.584037 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kwqwl_f89b68c7-da6e-447c-82f8-8662fe8efd35/registry-server/0.log" Dec 03 19:32:39 crc kubenswrapper[4787]: I1203 19:32:39.766066 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-utilities/0.log" Dec 03 19:32:39 crc kubenswrapper[4787]: I1203 19:32:39.772510 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-content/0.log" Dec 03 19:32:39 crc kubenswrapper[4787]: I1203 19:32:39.892594 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-content/0.log" Dec 03 19:32:40 crc kubenswrapper[4787]: I1203 19:32:40.066528 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-utilities/0.log" Dec 03 19:32:40 crc kubenswrapper[4787]: I1203 19:32:40.086271 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/extract-content/0.log" Dec 03 19:32:40 crc kubenswrapper[4787]: I1203 19:32:40.275436 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-utilities/0.log" Dec 03 19:32:40 crc kubenswrapper[4787]: I1203 19:32:40.411948 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-7cj8s_280abffe-0dc9-4096-8574-74694e459f31/registry-server/0.log" Dec 03 19:32:40 crc kubenswrapper[4787]: I1203 19:32:40.476895 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-5mwph_4ffe2b99-3c93-493c-899f-be8d2e6a65b7/registry-server/0.log" Dec 03 19:32:40 crc kubenswrapper[4787]: I1203 19:32:40.510644 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-utilities/0.log" Dec 03 19:32:40 crc kubenswrapper[4787]: I1203 19:32:40.526900 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-content/0.log" Dec 03 19:32:40 crc kubenswrapper[4787]: I1203 19:32:40.589005 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-content/0.log" Dec 03 19:32:40 crc kubenswrapper[4787]: I1203 19:32:40.743763 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-utilities/0.log" Dec 03 19:32:40 crc kubenswrapper[4787]: I1203 19:32:40.771776 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/extract-content/0.log" Dec 03 19:32:41 crc kubenswrapper[4787]: I1203 19:32:41.135810 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-j95ht_3efdf95a-b018-4550-9853-61869474de08/registry-server/0.log" Dec 03 19:32:46 crc kubenswrapper[4787]: I1203 19:32:46.844619 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:32:46 crc kubenswrapper[4787]: E1203 19:32:46.845426 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:32:54 crc kubenswrapper[4787]: I1203 19:32:54.121374 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-668cf9dfbb-84bqx_d696b783-ca2e-495e-83fe-d33ab80fdf3f/prometheus-operator/0.log" Dec 03 19:32:54 crc kubenswrapper[4787]: I1203 19:32:54.293667 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5c564fcc7c-569n9_1801fa2e-073e-486f-bd71-6abab103dd9f/prometheus-operator-admission-webhook/0.log" Dec 03 19:32:54 crc kubenswrapper[4787]: I1203 19:32:54.335614 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-5c564fcc7c-7zjj9_ede219fc-7a66-484b-831f-6b242694494c/prometheus-operator-admission-webhook/0.log" Dec 03 19:32:54 crc kubenswrapper[4787]: I1203 19:32:54.498536 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-d8bb48f5d-ppms6_bf31c2a1-79e3-47bd-8e8c-f9c6ce432f87/operator/0.log" Dec 03 19:32:54 crc kubenswrapper[4787]: I1203 19:32:54.541805 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-ui-dashboards-7d5fb4cbfb-qwtfr_c72c745a-9a1c-4374-b1a7-97279603edf7/observability-ui-dashboards/0.log" Dec 03 19:32:54 crc kubenswrapper[4787]: I1203 19:32:54.690586 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5446b9c989-6cfhr_3651062d-0769-4da5-bb95-7c5987fc2b7e/perses-operator/0.log" Dec 03 19:32:59 crc kubenswrapper[4787]: I1203 19:32:59.765988 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:32:59 crc kubenswrapper[4787]: E1203 19:32:59.767129 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:33:08 crc kubenswrapper[4787]: I1203 19:33:08.874195 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5cd867f87d-zwxll_c73f5b31-0cde-406e-800b-153981eeead6/kube-rbac-proxy/0.log" Dec 03 19:33:08 crc kubenswrapper[4787]: I1203 19:33:08.902541 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators-redhat_loki-operator-controller-manager-5cd867f87d-zwxll_c73f5b31-0cde-406e-800b-153981eeead6/manager/0.log" Dec 03 19:33:11 crc kubenswrapper[4787]: I1203 19:33:11.765962 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:33:11 crc kubenswrapper[4787]: E1203 19:33:11.766787 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:33:25 crc kubenswrapper[4787]: I1203 19:33:25.766231 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:33:25 crc kubenswrapper[4787]: E1203 19:33:25.766901 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:33:39 crc kubenswrapper[4787]: I1203 19:33:39.768975 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:33:39 crc kubenswrapper[4787]: E1203 19:33:39.789174 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:33:50 crc kubenswrapper[4787]: I1203 19:33:50.767469 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:33:50 crc kubenswrapper[4787]: E1203 19:33:50.768528 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:34:01 crc kubenswrapper[4787]: I1203 19:34:01.766698 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:34:01 crc kubenswrapper[4787]: E1203 19:34:01.767676 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:34:14 crc kubenswrapper[4787]: I1203 19:34:14.766757 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:34:14 crc kubenswrapper[4787]: E1203 19:34:14.767646 4787 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-t882k_openshift-machine-config-operator(b6597ac6-3ab2-4d2f-b38e-896795a7773d)\"" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" Dec 03 19:34:29 crc kubenswrapper[4787]: I1203 19:34:29.769222 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:34:30 crc kubenswrapper[4787]: I1203 19:34:30.494772 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"6b3bcd6f172904b24df1dc5c97af5f6dc6bd779c0f3fc670404ef893a89a2104"} Dec 03 19:34:33 crc kubenswrapper[4787]: I1203 19:34:33.694667 4787 scope.go:117] "RemoveContainer" containerID="be2ec87382979b629ed27a341498a17487ecd4c8688e58671f490640106984f3" Dec 03 19:35:14 crc kubenswrapper[4787]: I1203 19:35:14.060806 4787 generic.go:334] "Generic (PLEG): container finished" podID="9b1decf2-f0ad-4149-a10d-874c34248069" containerID="c3abc84527b3f1cf3a11e4fd690d2e06468536c47b52769d3bfc102d9a145807" exitCode=0 Dec 03 19:35:14 crc kubenswrapper[4787]: I1203 19:35:14.060958 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-srnnw/must-gather-9xczg" event={"ID":"9b1decf2-f0ad-4149-a10d-874c34248069","Type":"ContainerDied","Data":"c3abc84527b3f1cf3a11e4fd690d2e06468536c47b52769d3bfc102d9a145807"} Dec 03 19:35:14 crc kubenswrapper[4787]: I1203 19:35:14.062100 4787 scope.go:117] "RemoveContainer" containerID="c3abc84527b3f1cf3a11e4fd690d2e06468536c47b52769d3bfc102d9a145807" Dec 03 19:35:14 crc kubenswrapper[4787]: I1203 19:35:14.693221 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-srnnw_must-gather-9xczg_9b1decf2-f0ad-4149-a10d-874c34248069/gather/0.log" Dec 03 19:35:27 crc kubenswrapper[4787]: I1203 19:35:27.462206 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-srnnw/must-gather-9xczg"] Dec 03 19:35:27 crc kubenswrapper[4787]: I1203 19:35:27.465154 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-srnnw/must-gather-9xczg" podUID="9b1decf2-f0ad-4149-a10d-874c34248069" containerName="copy" containerID="cri-o://36ed78655baaa75f6af15ff9948fea643bce9069835bcf63055438fd251a35b6" gracePeriod=2 Dec 03 19:35:27 crc kubenswrapper[4787]: I1203 19:35:27.490673 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-srnnw/must-gather-9xczg"] Dec 03 19:35:27 crc kubenswrapper[4787]: I1203 19:35:27.657422 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-srnnw_must-gather-9xczg_9b1decf2-f0ad-4149-a10d-874c34248069/copy/0.log" Dec 03 19:35:27 crc kubenswrapper[4787]: I1203 19:35:27.658595 4787 generic.go:334] "Generic (PLEG): container finished" podID="9b1decf2-f0ad-4149-a10d-874c34248069" containerID="36ed78655baaa75f6af15ff9948fea643bce9069835bcf63055438fd251a35b6" exitCode=143 Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.199302 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-srnnw_must-gather-9xczg_9b1decf2-f0ad-4149-a10d-874c34248069/copy/0.log" Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.200220 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/must-gather-9xczg" Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.324398 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9b1decf2-f0ad-4149-a10d-874c34248069-must-gather-output\") pod \"9b1decf2-f0ad-4149-a10d-874c34248069\" (UID: \"9b1decf2-f0ad-4149-a10d-874c34248069\") " Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.324480 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rj42q\" (UniqueName: \"kubernetes.io/projected/9b1decf2-f0ad-4149-a10d-874c34248069-kube-api-access-rj42q\") pod \"9b1decf2-f0ad-4149-a10d-874c34248069\" (UID: \"9b1decf2-f0ad-4149-a10d-874c34248069\") " Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.332083 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b1decf2-f0ad-4149-a10d-874c34248069-kube-api-access-rj42q" (OuterVolumeSpecName: "kube-api-access-rj42q") pod "9b1decf2-f0ad-4149-a10d-874c34248069" (UID: "9b1decf2-f0ad-4149-a10d-874c34248069"). InnerVolumeSpecName "kube-api-access-rj42q". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.427585 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rj42q\" (UniqueName: \"kubernetes.io/projected/9b1decf2-f0ad-4149-a10d-874c34248069-kube-api-access-rj42q\") on node \"crc\" DevicePath \"\"" Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.549603 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9b1decf2-f0ad-4149-a10d-874c34248069-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "9b1decf2-f0ad-4149-a10d-874c34248069" (UID: "9b1decf2-f0ad-4149-a10d-874c34248069"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.631929 4787 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/9b1decf2-f0ad-4149-a10d-874c34248069-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.671658 4787 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-srnnw_must-gather-9xczg_9b1decf2-f0ad-4149-a10d-874c34248069/copy/0.log" Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.672211 4787 scope.go:117] "RemoveContainer" containerID="36ed78655baaa75f6af15ff9948fea643bce9069835bcf63055438fd251a35b6" Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.672235 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-srnnw/must-gather-9xczg" Dec 03 19:35:28 crc kubenswrapper[4787]: I1203 19:35:28.690176 4787 scope.go:117] "RemoveContainer" containerID="c3abc84527b3f1cf3a11e4fd690d2e06468536c47b52769d3bfc102d9a145807" Dec 03 19:35:29 crc kubenswrapper[4787]: I1203 19:35:29.778826 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b1decf2-f0ad-4149-a10d-874c34248069" path="/var/lib/kubelet/pods/9b1decf2-f0ad-4149-a10d-874c34248069/volumes" Dec 03 19:35:33 crc kubenswrapper[4787]: I1203 19:35:33.778783 4787 scope.go:117] "RemoveContainer" containerID="f9411b2119998f21f1280fc69e0b084d37bf3a215347b9d47cc0baaa83ced7f9" Dec 03 19:36:48 crc kubenswrapper[4787]: I1203 19:36:48.990442 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:36:48 crc kubenswrapper[4787]: I1203 19:36:48.991563 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:37:17 crc kubenswrapper[4787]: I1203 19:37:17.976948 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mtx97"] Dec 03 19:37:17 crc kubenswrapper[4787]: E1203 19:37:17.978338 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b1decf2-f0ad-4149-a10d-874c34248069" containerName="gather" Dec 03 19:37:17 crc kubenswrapper[4787]: I1203 19:37:17.978357 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b1decf2-f0ad-4149-a10d-874c34248069" containerName="gather" Dec 03 19:37:17 crc kubenswrapper[4787]: E1203 19:37:17.978380 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17e1369-d28b-4930-8d43-6f17d9289a7d" containerName="extract-utilities" Dec 03 19:37:17 crc kubenswrapper[4787]: I1203 19:37:17.978389 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17e1369-d28b-4930-8d43-6f17d9289a7d" containerName="extract-utilities" Dec 03 19:37:17 crc kubenswrapper[4787]: E1203 19:37:17.978405 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17e1369-d28b-4930-8d43-6f17d9289a7d" containerName="registry-server" Dec 03 19:37:17 crc kubenswrapper[4787]: I1203 19:37:17.978414 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17e1369-d28b-4930-8d43-6f17d9289a7d" containerName="registry-server" Dec 03 19:37:17 crc kubenswrapper[4787]: E1203 19:37:17.978431 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b1decf2-f0ad-4149-a10d-874c34248069" containerName="copy" Dec 03 19:37:17 crc kubenswrapper[4787]: I1203 19:37:17.978439 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b1decf2-f0ad-4149-a10d-874c34248069" containerName="copy" Dec 03 19:37:17 crc kubenswrapper[4787]: E1203 19:37:17.978488 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f17e1369-d28b-4930-8d43-6f17d9289a7d" containerName="extract-content" Dec 03 19:37:17 crc kubenswrapper[4787]: I1203 19:37:17.978497 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="f17e1369-d28b-4930-8d43-6f17d9289a7d" containerName="extract-content" Dec 03 19:37:17 crc kubenswrapper[4787]: I1203 19:37:17.978851 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="f17e1369-d28b-4930-8d43-6f17d9289a7d" containerName="registry-server" Dec 03 19:37:17 crc kubenswrapper[4787]: I1203 19:37:17.978870 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b1decf2-f0ad-4149-a10d-874c34248069" containerName="copy" Dec 03 19:37:17 crc kubenswrapper[4787]: I1203 19:37:17.978903 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b1decf2-f0ad-4149-a10d-874c34248069" containerName="gather" Dec 03 19:37:17 crc kubenswrapper[4787]: I1203 19:37:17.980975 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.003588 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mtx97"] Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.115139 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-utilities\") pod \"redhat-operators-mtx97\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.115236 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p7pws\" (UniqueName: \"kubernetes.io/projected/44de4938-6bce-4098-a549-2cb4141a6e5a-kube-api-access-p7pws\") pod \"redhat-operators-mtx97\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.115339 4787 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-catalog-content\") pod \"redhat-operators-mtx97\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.217503 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-catalog-content\") pod \"redhat-operators-mtx97\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.217717 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-utilities\") pod \"redhat-operators-mtx97\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.217784 4787 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p7pws\" (UniqueName: \"kubernetes.io/projected/44de4938-6bce-4098-a549-2cb4141a6e5a-kube-api-access-p7pws\") pod \"redhat-operators-mtx97\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.218231 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-catalog-content\") pod \"redhat-operators-mtx97\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.218323 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-utilities\") pod \"redhat-operators-mtx97\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.252532 4787 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p7pws\" (UniqueName: \"kubernetes.io/projected/44de4938-6bce-4098-a549-2cb4141a6e5a-kube-api-access-p7pws\") pod \"redhat-operators-mtx97\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.320491 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.889045 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mtx97"] Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.989574 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:37:18 crc kubenswrapper[4787]: I1203 19:37:18.989907 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:37:19 crc kubenswrapper[4787]: I1203 19:37:19.042540 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtx97" event={"ID":"44de4938-6bce-4098-a549-2cb4141a6e5a","Type":"ContainerStarted","Data":"c34b78c54330252b6c61ab5c66d7d8d69f5cea460e8a40ffdc50802ce0b78a22"} Dec 03 19:37:20 crc kubenswrapper[4787]: I1203 19:37:20.061654 4787 generic.go:334] "Generic (PLEG): container finished" podID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerID="319a4b92c2f57c50a84a64d37895bcddf095b83e90b6ded8684676c21676b066" exitCode=0 Dec 03 19:37:20 crc kubenswrapper[4787]: I1203 19:37:20.061729 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtx97" event={"ID":"44de4938-6bce-4098-a549-2cb4141a6e5a","Type":"ContainerDied","Data":"319a4b92c2f57c50a84a64d37895bcddf095b83e90b6ded8684676c21676b066"} Dec 03 19:37:20 crc kubenswrapper[4787]: I1203 19:37:20.069113 4787 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 03 19:37:21 crc kubenswrapper[4787]: I1203 19:37:21.073005 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtx97" event={"ID":"44de4938-6bce-4098-a549-2cb4141a6e5a","Type":"ContainerStarted","Data":"aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46"} Dec 03 19:37:25 crc kubenswrapper[4787]: I1203 19:37:25.131966 4787 generic.go:334] "Generic (PLEG): container finished" podID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerID="aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46" exitCode=0 Dec 03 19:37:25 crc kubenswrapper[4787]: I1203 19:37:25.132092 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtx97" event={"ID":"44de4938-6bce-4098-a549-2cb4141a6e5a","Type":"ContainerDied","Data":"aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46"} Dec 03 19:37:26 crc kubenswrapper[4787]: I1203 19:37:26.147718 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtx97" event={"ID":"44de4938-6bce-4098-a549-2cb4141a6e5a","Type":"ContainerStarted","Data":"dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3"} Dec 03 19:37:26 crc kubenswrapper[4787]: I1203 19:37:26.171486 4787 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mtx97" podStartSLOduration=3.715026179 podStartE2EDuration="9.171463053s" podCreationTimestamp="2025-12-03 19:37:17 +0000 UTC" firstStartedPulling="2025-12-03 19:37:20.065800005 +0000 UTC m=+8656.883270974" lastFinishedPulling="2025-12-03 19:37:25.522236869 +0000 UTC m=+8662.339707848" observedRunningTime="2025-12-03 19:37:26.170398655 +0000 UTC m=+8662.987869634" watchObservedRunningTime="2025-12-03 19:37:26.171463053 +0000 UTC m=+8662.988934012" Dec 03 19:37:28 crc kubenswrapper[4787]: I1203 19:37:28.320966 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:28 crc kubenswrapper[4787]: I1203 19:37:28.321731 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:29 crc kubenswrapper[4787]: I1203 19:37:29.400202 4787 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mtx97" podUID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerName="registry-server" probeResult="failure" output=< Dec 03 19:37:29 crc kubenswrapper[4787]: timeout: failed to connect service ":50051" within 1s Dec 03 19:37:29 crc kubenswrapper[4787]: > Dec 03 19:37:38 crc kubenswrapper[4787]: I1203 19:37:38.406966 4787 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:38 crc kubenswrapper[4787]: I1203 19:37:38.476970 4787 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:38 crc kubenswrapper[4787]: I1203 19:37:38.677491 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mtx97"] Dec 03 19:37:40 crc kubenswrapper[4787]: I1203 19:37:40.310389 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mtx97" podUID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerName="registry-server" containerID="cri-o://dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3" gracePeriod=2 Dec 03 19:37:40 crc kubenswrapper[4787]: I1203 19:37:40.923531 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.051701 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-utilities\") pod \"44de4938-6bce-4098-a549-2cb4141a6e5a\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.051814 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-catalog-content\") pod \"44de4938-6bce-4098-a549-2cb4141a6e5a\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.051878 4787 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p7pws\" (UniqueName: \"kubernetes.io/projected/44de4938-6bce-4098-a549-2cb4141a6e5a-kube-api-access-p7pws\") pod \"44de4938-6bce-4098-a549-2cb4141a6e5a\" (UID: \"44de4938-6bce-4098-a549-2cb4141a6e5a\") " Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.052974 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-utilities" (OuterVolumeSpecName: "utilities") pod "44de4938-6bce-4098-a549-2cb4141a6e5a" (UID: "44de4938-6bce-4098-a549-2cb4141a6e5a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.067481 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44de4938-6bce-4098-a549-2cb4141a6e5a-kube-api-access-p7pws" (OuterVolumeSpecName: "kube-api-access-p7pws") pod "44de4938-6bce-4098-a549-2cb4141a6e5a" (UID: "44de4938-6bce-4098-a549-2cb4141a6e5a"). InnerVolumeSpecName "kube-api-access-p7pws". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.154254 4787 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-utilities\") on node \"crc\" DevicePath \"\"" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.154503 4787 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p7pws\" (UniqueName: \"kubernetes.io/projected/44de4938-6bce-4098-a549-2cb4141a6e5a-kube-api-access-p7pws\") on node \"crc\" DevicePath \"\"" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.174425 4787 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44de4938-6bce-4098-a549-2cb4141a6e5a" (UID: "44de4938-6bce-4098-a549-2cb4141a6e5a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.261337 4787 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44de4938-6bce-4098-a549-2cb4141a6e5a-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.322838 4787 generic.go:334] "Generic (PLEG): container finished" podID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerID="dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3" exitCode=0 Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.322880 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtx97" event={"ID":"44de4938-6bce-4098-a549-2cb4141a6e5a","Type":"ContainerDied","Data":"dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3"} Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.322924 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mtx97" event={"ID":"44de4938-6bce-4098-a549-2cb4141a6e5a","Type":"ContainerDied","Data":"c34b78c54330252b6c61ab5c66d7d8d69f5cea460e8a40ffdc50802ce0b78a22"} Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.322939 4787 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mtx97" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.322947 4787 scope.go:117] "RemoveContainer" containerID="dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.353957 4787 scope.go:117] "RemoveContainer" containerID="aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.388484 4787 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mtx97"] Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.389392 4787 scope.go:117] "RemoveContainer" containerID="319a4b92c2f57c50a84a64d37895bcddf095b83e90b6ded8684676c21676b066" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.404180 4787 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mtx97"] Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.455826 4787 scope.go:117] "RemoveContainer" containerID="dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3" Dec 03 19:37:41 crc kubenswrapper[4787]: E1203 19:37:41.456401 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3\": container with ID starting with dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3 not found: ID does not exist" containerID="dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.456469 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3"} err="failed to get container status \"dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3\": rpc error: code = NotFound desc = could not find container \"dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3\": container with ID starting with dc2aad91a6bd03db5e5824ae675e649affea991e497bc69973745661644b5bd3 not found: ID does not exist" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.456497 4787 scope.go:117] "RemoveContainer" containerID="aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46" Dec 03 19:37:41 crc kubenswrapper[4787]: E1203 19:37:41.456950 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46\": container with ID starting with aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46 not found: ID does not exist" containerID="aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.456980 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46"} err="failed to get container status \"aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46\": rpc error: code = NotFound desc = could not find container \"aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46\": container with ID starting with aae7beb8677b2bade7847a38f1db0d38145de19b56895b56b51941e004f34f46 not found: ID does not exist" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.456999 4787 scope.go:117] "RemoveContainer" containerID="319a4b92c2f57c50a84a64d37895bcddf095b83e90b6ded8684676c21676b066" Dec 03 19:37:41 crc kubenswrapper[4787]: E1203 19:37:41.457376 4787 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"319a4b92c2f57c50a84a64d37895bcddf095b83e90b6ded8684676c21676b066\": container with ID starting with 319a4b92c2f57c50a84a64d37895bcddf095b83e90b6ded8684676c21676b066 not found: ID does not exist" containerID="319a4b92c2f57c50a84a64d37895bcddf095b83e90b6ded8684676c21676b066" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.457395 4787 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"319a4b92c2f57c50a84a64d37895bcddf095b83e90b6ded8684676c21676b066"} err="failed to get container status \"319a4b92c2f57c50a84a64d37895bcddf095b83e90b6ded8684676c21676b066\": rpc error: code = NotFound desc = could not find container \"319a4b92c2f57c50a84a64d37895bcddf095b83e90b6ded8684676c21676b066\": container with ID starting with 319a4b92c2f57c50a84a64d37895bcddf095b83e90b6ded8684676c21676b066 not found: ID does not exist" Dec 03 19:37:41 crc kubenswrapper[4787]: I1203 19:37:41.776308 4787 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44de4938-6bce-4098-a549-2cb4141a6e5a" path="/var/lib/kubelet/pods/44de4938-6bce-4098-a549-2cb4141a6e5a/volumes" Dec 03 19:37:48 crc kubenswrapper[4787]: I1203 19:37:48.989578 4787 patch_prober.go:28] interesting pod/machine-config-daemon-t882k container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 03 19:37:48 crc kubenswrapper[4787]: I1203 19:37:48.990541 4787 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 03 19:37:48 crc kubenswrapper[4787]: I1203 19:37:48.990629 4787 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-t882k" Dec 03 19:37:48 crc kubenswrapper[4787]: I1203 19:37:48.992448 4787 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6b3bcd6f172904b24df1dc5c97af5f6dc6bd779c0f3fc670404ef893a89a2104"} pod="openshift-machine-config-operator/machine-config-daemon-t882k" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 03 19:37:48 crc kubenswrapper[4787]: I1203 19:37:48.992618 4787 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-t882k" podUID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerName="machine-config-daemon" containerID="cri-o://6b3bcd6f172904b24df1dc5c97af5f6dc6bd779c0f3fc670404ef893a89a2104" gracePeriod=600 Dec 03 19:37:49 crc kubenswrapper[4787]: I1203 19:37:49.429353 4787 generic.go:334] "Generic (PLEG): container finished" podID="b6597ac6-3ab2-4d2f-b38e-896795a7773d" containerID="6b3bcd6f172904b24df1dc5c97af5f6dc6bd779c0f3fc670404ef893a89a2104" exitCode=0 Dec 03 19:37:49 crc kubenswrapper[4787]: I1203 19:37:49.429451 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerDied","Data":"6b3bcd6f172904b24df1dc5c97af5f6dc6bd779c0f3fc670404ef893a89a2104"} Dec 03 19:37:49 crc kubenswrapper[4787]: I1203 19:37:49.430165 4787 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-t882k" event={"ID":"b6597ac6-3ab2-4d2f-b38e-896795a7773d","Type":"ContainerStarted","Data":"b4df71267a1747ed03be9d68cb41d0fa9390bb3081259fde6c557e8a1b2d4a7c"} Dec 03 19:37:49 crc kubenswrapper[4787]: I1203 19:37:49.430192 4787 scope.go:117] "RemoveContainer" containerID="c358a1c2749e038f148487f56a3bb99825ef6215dde42eb874a29a1a5e16f294" Dec 03 19:39:29 crc kubenswrapper[4787]: I1203 19:39:29.479160 4787 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-g5jrv"] Dec 03 19:39:29 crc kubenswrapper[4787]: E1203 19:39:29.481625 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerName="extract-utilities" Dec 03 19:39:29 crc kubenswrapper[4787]: I1203 19:39:29.481640 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerName="extract-utilities" Dec 03 19:39:29 crc kubenswrapper[4787]: E1203 19:39:29.481652 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerName="registry-server" Dec 03 19:39:29 crc kubenswrapper[4787]: I1203 19:39:29.481659 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerName="registry-server" Dec 03 19:39:29 crc kubenswrapper[4787]: E1203 19:39:29.481690 4787 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerName="extract-content" Dec 03 19:39:29 crc kubenswrapper[4787]: I1203 19:39:29.481698 4787 state_mem.go:107] "Deleted CPUSet assignment" podUID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerName="extract-content" Dec 03 19:39:29 crc kubenswrapper[4787]: I1203 19:39:29.481937 4787 memory_manager.go:354] "RemoveStaleState removing state" podUID="44de4938-6bce-4098-a549-2cb4141a6e5a" containerName="registry-server" Dec 03 19:39:29 crc kubenswrapper[4787]: I1203 19:39:29.483468 4787 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-g5jrv" Dec 03 19:39:29 crc kubenswrapper[4787]: I1203 19:39:29.512852 4787 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-g5jrv"] var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515114110767024450 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015114110770017357 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015114067307016511 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015114067310015453 5ustar corecore